We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 5
Iaas provider’s access to fundamental resources such as physical PaaS is a cloud computing service model that provides a platform
ovides a platform and environment to allow
machines, virtual machines, virtual storage, etc., Apart from these developers to build, deploy and manage applications without the complexity of managing resources, the IaaS also offers: • Virtual machine disk storage • Virtual local underlying infrastructure (eg. servers, storage, networking). It abstracts away much of the area network (VLANs) • Load balancers • IP addresses • Software bundles technical complexity involved in deploying and maintaining infrastructure. Key Features • Characteristics: Here are the characteristics of IaaS service model: • Virtual Development Frameworks: PaaS provides development tools, libraries & frameworks (eg, for web, mobile applications). • Managed Infrastructure: Users don’t need to manage hardware, machines with pre-installed software. • Virtual machines with pre-installed networking, or storage. The cloud provider manages the infrastructure, allowing developers Operating Systems such as Windows, Linux, and Solaris. • On-demand to focus on writing code.• Scalability: PaaS platforms typically provide automatic scaling, availability of resources. • Allows to store copies of particular data in meaning applications can scale up or down based on traffic or demand. •Integration Services: different locations. • The computing resources can be easily scaled up and Many PaaS offerings include APIs, databases, and messaging queues for seamless integration down. Benefits: IaaS allows the cloud provider to freely locate the between different services. •DevOps Integration: PaaS often includes CI/CD(Continuous infrastructure over the Internet in a cost effective manner. Some of the Integration and Continuous Deployment)tools for automating the software development key benefits of IaaS are listed below: • Full Control of the computing lifecycle. Eg. GoGrid CloudCenter, Google AppEngine, Windows Azure Platform, IBM Cloud resources through Administrative Access to VMs. • Flexible and Efficient Foundry, Heroku, Red Hat OpenShift renting of Computer Hardware. • Portability, Interoperability with Legacy Benefits: • Lower Administrative Overhead Consumer need not to bother much about the Applications. Eg. Amazon Elastic Compute Cloud (EC2), Eucalyptus, GoGrid, administration because it's the responsibility of cloud provider. • Lower Total Cost Of Ownership Consumer need not purchase expensive hardware, servers, power and data FlexiScale, Linode, RackSpace Cloud (PTO) storage. • Scalable Solutions It is very easy to scale up or down automatically based on application resource demands. • More Current System Softwar It is the responsibility of the The Cloud Cube Model: The Jericho Forum has identified four criteria to cloud provider to maintain software versions and patch installations. Issues: • Lack Of Portability Between Paas Clouds Although standard languages are used yet the differentiate cloud formations from each other and the manner of their implementations of platforms services may vary. For example, file, queue, or hash table provision. The Cloud Cube Model effectively summarizes these four interfaces of one platform may differ from another, making it difficult to transfer workloads dimensions: 1. Internal/External 2. Proprietary/Open 3. Perimeterised/De- from one platform to another. • Event Based Processor Scheduling perimeterized Architectures 4. Insourced/Outsourced. Dimension 1: Internal/External This dimension defines the physical location of the data; where does the cloud form exist -inside or outside organization boundaries? Public Cloud: In a public cloud, the infrastructure and services are owned and operated by a If the cloud form is within the organization's physical boundaries, then it is third-party provider and shared among multiple customers (tenants). The resources (such as internal. If it is outside the organization's physical boundaries, then it' is servers, storage, and networking) are delivered over the internet. Benefits: • Cost Effective: external. It's important to note that the assumption that internal is necessarily more secure than external is false. The most secure usage model Since public cloud share same resources with large number of consumer, it has low cost. • is the effective use of both internal and external cloud forms. Dimension 2: Reliability: Since public cloud employs large number of resources from different locations, if Proprietary/Open This dimension defines the state of ownership of the cloud any of the resource fail, public cloud can employ another one. • Flexibility It is also very easy technology, services, interfaces, etc. It indicates the degree of to integrate public cloud with private cloud, hence gives consumers a flexible approach. • interoperability, as well as enabling data/application transportability Location Independence Since, public cloud services are delivered through Internet, therefore between an organization's own systems and other cloud forms and the ability ensures location independence. • Utility Style Costing Public cloud is also based on pay-per- to withdraw the data from a cloud form, or to move it to another without use model and resources are accessible whenever consumer needs it. • High Scalability Cloud constraint. This dimension indicates any constraints on being able to share resources are made available on demand from a pool of resources, i.e., they can be scaled up apps. "Proprietary" suggests that the organization providing the service is or down according the requirement. Disadvantages: Here are the disadvantages of public keeping the means of provision under its ownership. By contrast, "open" cloud model: • Low Security In public cloud model, data is hosted off-site and resources are clouds use technology that is not proprietary, which means that there are shared publicly, therefore does not ensure higher level of security. • Less Customizable It is likely to be more suppliers, and the organization is not as constrained in terms comparatively less customizable than private cloud. Hybrid Cloud: A hybrid cloud combines of ability to share data and collaborate with selected parties. Experts suggest both public and private cloud infrastructures. It allows data and applications to be shared that open clouds most effectively enhance collaboration between multiple between them, providing greater flexibility and optimization by using both cloud types. organizations. Dimension 3: Perimeterised/De-perimeterised Architectures Benefits: • Scalability: It offers both features of public cloud scalability and private cloud Perimeterised" suggest a system that continues to operate within the scalability. • Flexibility: It offers both secure resources and scalable public resources. • Cost traditional IT perimeter, often characterized by "network firewalls." This Efficiencies: Public cloud are more cost effective than private, therefore hybrid cloud can approach is known to inhibit collaboration. (PTO) have this saving. • Security: Private cloud in hybrid cloud ensures higher degree of security. Disadvantages: • Networking Issues: Networking becomes complex due to presence of Simple Storage Service (S3) It is a flexible, scalable, and highly available private and public cloud. • Security Compliance: It is necessary to ensure that cloud services storage web service. It is like having an infinitely large hard drive where users are compliant with organization's security policies. can store files which are then accessible via a unique URL. S3 also supports access control, expiration times, and several other useful features. AWS is an expanded cloud computing platform provided by Amazon Company. AWS Additionally, the payment model for S3 is "pay as you go" so you'll only be provides a wide range of services with a pay-as-per-use pricing model over the Internet such billed for the amount of data users store and how much bandwidth users use as Storage, Computing power, Databases, Machine Learning services, and much more. AWS to transfer it in and out. Amazon S3 is used for various purposes in the Cloud facilitates for both businesses and individual users with effectively hosting the applications, because of its robust features with scaling and Securing of data. It helps storing the data securely, and making use of a wide variety of tools and services improving people with all kinds of use cases from fields such as Mobile/Web management flexibility for IT resources. AWS comes up with its own network infrastructure applications, Big data, Machine Learning and many more. Amazon S3 bucket on establishing the datacenters in different regions mostly all over the world. Its global Infrastructure acts as a backbone for operations and services provided by AWS. It facilitates is a fundamental Storage Container feature in AWS S3 Service. It provides a the users on creating secure environments using Amazon VPCs(Virtual Private Clouds). secure and scalable repository for storing of Objects such as Text data, Essential services like Amazon EC2 and Amazon S3 for utilizing the compute and storage Images, Audio and Video files over AWS Cloud. Each S3 bucket name should service with elastic scaling. It supports the dynamic scaling of the applications with the be named globally unique and should be configured with ACL (Access Control services such as Auto Scaling and Elastic Load Balancing ( AWS ELB ). It provides a good user- List). Uses: 1.Data Storage: Amazon s3 acts as the best option for scaling both friendly AWS Management Console facilitating seamless configuration and management of small and large storage applications. It helps in storing and retrieving the AWS services to the Users. Its Architecture ensures high availability , fault tolerance making data-intensive applications as per needs in ideal time. 2.Backup and AWS a versatile powerful Cloud Computing Platform. Recovery: Many Organizations are using Amazon S3 to backup their critical Features:1.Cost-Effective: i. AWS eliminates upfront costs, long-term commitments, or data and maintain the data durability and availability for recovery needs. minimum spending requirements.ii. Users can scale resources up or down based on demand, 3.Hosting Static Websites: Amazon S3 facilitates in storing HTML, CSS and paying only for what they use.2.Scalable and Elastic: i. Scalability enables automatic other web content from Users/developers allowing them for hosting Static adjustments to computing capacity as demand increases or decreases.ii. Elasticity ensures Websites benefiting with low-latency access and cost-effectiveness. (PTO) efficient traffic distribution across resources like EC2 instances, containers, IP addresses, and Lambda functions.3.Security and Compliance:i. AWS ensures robust data security with rules Hypervisor: A low-level program is required to provide system resource that filter traffic by type, protocol, port range, and source.ii. Security groups protect EC2 access to virtual machines, and this program is referred to as the hypervisor instances at both protocol and port access levels.4.Serverless Cloud Functions: i. AWS allows or Virtual Machine Monitor (VMM). There are different types of virtual developers to focus on building applications by managing the underlying infrastructure.ii. It machines. They are: Type 1 VM or native VM. Examples of Type 1 Virtual handles scaling, patching, and administration, ensuring smooth backend operations without Machine Monitors are LynxSecure, RTS Hypervisor, Oracle VM, Sun xVM server management.5.Flexibility: i. Users can choose their operating system, programming Server, VirtualLogix VLX, VMware ESX and ESXi, and Wind River VxWorks, language (Java, Python, Node.js, etc.), database, and content delivery location.ii. AWS among others. The operating system loaded into a virtual machine is referred supports hybrid application models, allowing some components to run in data centers while to as the guest operating system, and there is no constraint on running the others operate in the cloud.6.Experienced: i. AWS continually improves its infrastructure, same guest on multiple VMs on a physical system. Type 1 VMs have no host leveraging over 16 years of experience in cloud management.ii. It serves millions globally and operating system because they are installed on a bare system. Type 2 or has evolved since 2006, supporting a vast and diverse customer base. hosted VM. Examples of Type 2 Virtual Machine Monitors are Containers, KVM, Microsoft Hyper V, Parallels Desktop for Mac, Wind River Simics, Elastic Compute Cloud It is the central piece of the AWS ecosystem. EC2 provides flexible, on- VMWare Fusion, Virtual Server 2005 R2, Xen, Windows Virtual PC, and demand computing resources with a "pay as you go" pricing model. You can deploy your VMware Workstation 6.0 and Server, among others. This is a very rich product category. Type 2 virtual machines are installed over a host operating system; applications in EC2 servers without any worrying about the underlying infrastructure. You for Microsoft Hyper-V, that operating system would be Windows Server. On configure the EC2-Instance in a very secure manner by using the VPC, Subnets, and Security a Type 2 VM, a software interface is created that emulates the devices with groups. You can scale up and scale down the instance based on the incoming traffic of the which a system would normally interact. application. The AWS EC2 Instance types are as follows: General Purpose Instances, Compute Virtualization is a technique, which allows sharing single physical instance of Optimized Instances, Memory-Optimized Instances, Storage Optimized Instances, Accelerated an application or resource among multiple organizations or tenants Computing Instances. Features of AWS EC2: 1. Functionality: Amazon EC2 provides a virtual (customers). It does so by assigning a logical name to a physical resource and computing platform where users can perform various operations, launch new instances, and providing a pointer to that physical resource when demanded. Virtualization customize environments as needed. It offers default AMIs with pre-configured resources like Concept: Creating a virtual machine over existing operating system and RAM, ROM, storage, etc. and supports user-defined AMIs combining custom and default hardware is referred as Hardware Virtualization. Virtual Machines provide an settings. These custom AMIs can be stored for future use, simplifying the creation of new environment that is logically separated from the underlying hardware. The instances without reconfiguration. 2. Operating Systems: Amazon EC2 includes a wide range machine on which the virtual machine is created is known as host machine of operating systems to choose from while selecting your AMI. Not only are these selected and virtual machine is referred as a guest machine. This virtual machine is options, but users are also even given the privilege to upload their own operating systems and managed by a software or firmware, which is known as hypervisor. opt for that while selecting AMI during launching an EC2 instance. Currently, AWS has the Characteristics: •Partitioning: In virtualization, many applications and following most preferred set of operating systems available on the EC2 console: Amazon Linux, operating systems (OSes) are supported in a single physical system by partitioning (separating) the available resources. •Isolation: Each virtual Windows Server, Ubuntu Server, SUSE Linux, Red Hat Linux. 3. Software: Amazon is single- machine is isolated from its host physical system and other virtualized handedly ruling the cloud computing market, because of the variety of options available on machines. Because of this isolation, if one virtual-instance crashes, it doesn't EC2 for its users. It allows its users to choose from various software present to run on their EC2 affect the other virtual machines. In addition, data isn't shared between one machines. This whole service is allocated to AWS Marketplace on the AWS platform. Numerous virtual container and another. •Encapsulation: A virtual machine can be software like SAP, LAMP, Drupal, etc are available on AWS to use. 4.Scalability and Reliability: represented (and even stored) as a single file, so you can identify it easily EC2 provides us the facility to scale up or scale down as per the needs. All dynamic scenarios based on the service it provides. In essence, the encapsulated process could can be easily tackled by EC2 with the help of this feature. And because of the flexibility of be a business service. This encapsulated virtual machine can be presented to volumes and snapshots, it is highly reliable for its users. Due to the scalable nature of the an application as a complete entity. Therefore, encapsulation can protect machine, many organizations like Flipkart, and Amazon rely on these days whenever each application so that it doesn't interfere with another application. humongous traffic occurs on their portals. Paravirtualization requires that the host operating system provide a virtual machine interface for the guest operating system and that the guest access hardware through that host VM. An operating system running as a guest on a paravirtualization system must be ported to work with the host interface. (PTO) SaaS is a cloud computing model that delivers software applications over the internet. SaaS …Issues: • Compatibility With Legacy Security Vulnerabilities: Because providers host and maintain the software, while users access it via a web browser or API. In IaaS offers the consumer to run legacy software in provider's this model, users don’t need to worry about the underlying infrastructure, operating systems, infrastructure, therefore it exposes consumers to all of the security or software updates. SaaS applications, some of them are listed below: • Billing and Invoicing vulnerabilities of such legacy software. • Virtual Machine Sprawl: The VM System • Customer Relationship Management (CRM) applications • Help Desk Applications • can become out of date with respect to security updates because IaaS Human Resource (HR) Solutions. Key Features • On-demand Access: Users can access allows the consumer to operate the virtual machines in running, applications over the internet without installing or maintaining them locally. • Subscription suspended and off state. However, the provider can automatically Model: SaaS services are often offered on a subscription basis (e.g., monthly or yearly) or pay- as-you-go pricing. • Automatic Updates and Maintenance: The SaaS provider handles software update such VMs, but this mechanism is hard and complex. • Robustness updates, security patches, and system maintenance. • Multi-tenancy: A single instance of the Of Vm-Level Isolation: IaaS offers an isolated environment to individual software serves multiple customers (tenants), though each user’s data is isolated. • Scalability: consumers through hypervisor. Hypervisor is a software layer that SaaS applications can scale based on the number of users or required resources. Eg : Google includes hardware support for virtualization to split a physical computer suite, microsoft 365, salesforce.com, oracle on demand, sql azure, dropbox, zoom. into multiple virtual machines. • Data Erase Practices: The consumer Benefits: • Modest Software Tools • Efficient use of Software Licenses • Centralized uses virtual machines that in turn use the common disk resources Management & Data • Platform responsibilities managed by provider • Multitenant solutions provided by the cloud provider. When the consumer releases the Issues: • Browser Based Risks: If the consumer visits malicious website and browser becomes resource, the cloud provider must ensure that next consumer to rent the infected, and the subsequent access to SaaS application might compromise the consumer's resource does not observe data residue from previous consumer data. To avoid such risks, the consumer can use multiple browsers and dedicate a specific browser to access SaaS applications or can use virtual desktop while accessing the SaaS applications. • Network Dependence: The SaaS application can be delivered only when … Operating within such areas means extending an organization's perimeter network is continuously available. Also network should be reliable but the network reliability into the external cloud computing domain via a VPN and operating the virtual cannot be guaranteed either by cloud provider or the consumer. • Lack Of Portability Between server in its own IP domain. The organization uses its own directory services Saas Clouds: Transferring workloads from one SaaS cloud to another is not so easy because to control access. Once the computing task is complete, the perimeter is work flow, business logics, user interfaces, support scripts can be provider specific withdrawn to its original, traditional position. "De-perimeterised" suggests that the system perimeter is designed following the principles outlined in the Private Cloud • Description: A private cloud is a cloud infrastructure dedicated to a single Jericho Forum's Commandments and Collaboration Oriented Architectures organization. It can be hosted on-premises or by a third-party provider. The organization has Framework. De-perimeterised areas in the Cloud Cube Model use both internal and external domains, but the collaboration or sharing of data should more control over the security, privacy, and data management. Benefits: • Higher Security not be seen as internal or external. Rather, it is controlled by and limited to And Privacy: Private cloud operations are not available to general public and resources are the parties that the using organizations select. Dimension 4: shared from distinct pool of resources, therefore, ensures high security and privacy. • More Insourced/Outsourced This dimension has two states in each of the eight Control: Private clouds have more control on its resources and hardware than public cloud cloud forms. It responds to the question: who does the business want running because it is accessed only within an organization. • Cost And Energy Efficiency: Private cloud the clouds? "Outsourced" means that the service is provided by a third party. resources are not as cost effective as public clouds but they offer more efficiency than public Insourced means that the service is provided by the own staff under its cloud. Disadvantages: Here are the disadvantages of using private cloud model: • Restricted control. These states describe the party managing the delivery of the cloud Area: Private cloud is only accessible locally and is very difficult to deploy globally. • Inflexible service(s) used by the organization. Pricing: In order to fulfill demand, purchasing new hardware is very costly. • Limited Open SaaS combines SaaS delivery with open-source principles. The Scalability: Private cloud can be scaled only within capacity of internal hosted resources. software is open-source, allowing customers to modify, extend, or integrate Community Cloud • Description: A community cloud is shared by multiple organizations with the application according to their needs, while the cloud provider hosts and common interests (such as similar regulatory, security, or compliance requirements). It can maintains the platform. This allows businesses to customize SaaS be managed by the organizations or a third-party provider. Benefits: • Cost Effective: applications while benefiting from cloud infrastructure. Eg. WordPress, Community cloud offers same advantage as that of private cloud at low cost. • Sharing Nextcloud. Service-Oriented Architecture (SOA) is a design approach in Between Organizations: Community cloud provides an infrastructure to share cloud which software components (services) are loosely coupled and can interact resources and capabilities among several organizations. • Security: Community cloud is over a network. Each service provides specific functionality (e.g., payment comparatively more secure than the public cloud. Issues: • Since all data is housed at one processing, user authentication) and can be reused across different location, one must be careful in storing data in community cloud because it might be applications. In SaaS, SOA enables applications to be modular, scalable, and accessible by others. • It is also challenging to allocate responsibilities of governance, security integrated with other services or applications via APIs. and cost. ….4. Data Archiving: Amazon S3 Glacier service integration helps as a cost- effective solution for long-term data storing which are less frequently accessed applications. 5.Big Data Analytics: Amazon S3 is often considered Amazon AWS components: 1.Route53: It is a highly available, scalable, and feature rich domain as data lake because of its capacity to store large amounts of both structured name service (DNS) web service. What a DNS service does is translate a domain name like and unstructured data offering seamless integration with other AWS "setfive.com" into an IP address like 64.22.80.79 which allows a client's computer to "find" the correct server for a given domain name. In addition, Route53 also has several advanced features Analytics and AWS Machine Learning Services. Features of Amazon S3: normally only available in pricey enterprise DNS solutions. Route53 would typically replace the DNS 1.Durability: AWS claims Amazon S3 to have a 99.999999999% of durability service provided by common registrar like GoDaddy or Register.com. 2.Simple Email Service: It is a (11 9’s). This means the possibility of losing your data stored on S3 is one in hosted transactional email service. It allows uses to easily send highly deliverable emails using a a billion. 2.Availability: AWS ensures that the up-time of AWS S3 is 99.99% RESTful API call or via regular SMTP without running his or her own email infrastructure. 3.Identity for standard access. Note that availability is related to being able to access and Access Management: It provides enhanced security and identity management for users AWS data and durability is related to losing data altogether. 3.Server-Side- account. In addition, it allows them to enable "multi factor" authentication to enhance the security Encryption (SSE): AWS S3 supports three types of SSE models:SSE-S3: AWS of their AWS account. 4.S3, 5.EC2, 6.Elastic Block Storage It provides persist storage volumes that S3 manages encryption keys, SSE-C: The customer manages encryption keys, attach to EC2 instances to allow you to persist data past the lifespan of a single EC2. Due to the SSE-KMS: The AWS Key Management Service (KMS) manages the encryption architecture of elastic compute cloud. all the storage systems on an instance are ephemeral. This means that when an instance is terminated all the data stored on that instance is lost. EBS keys.4.File Size support: AWS S3 can hold files of size ranging from 0 bytes addresses this issue by providing persistent storage that appears on instances as a regular hard to 5 terabytes. A 5TB limit on file size should not be a blocker for most of the drive. Features of EBS: 1.Scalability: EBS volume sizes and features can be scaled as per the needs applications in the world. 5.Infinite storage space: Theoretically AWS S3 is of the system. This can be done in two ways: •Take a snapshot of the volume and create a new supposed to have infinite storage space. This makes S3 infinitely scalable for volume using the Snapshot with new updated features. •Updating the existing EBS volume from all kinds of use cases.6.Pay as you use: The users are charged according to the console. 2.Backup: Users can create snapshots of EBS volumes that act as backups.•Snapshot the S3 storage they hold. can be created manually at any point in time or can be scheduled. •Snapshots are stored on AWS S3 and are charged according to the S3 storage charges. •Snapshots are incremental in nature. •New volumes across regions can be created from snapshots. 3.Encryption: Encryption can be a ….In a Full virtualization scheme, the VM is installed as a Type 1 Hypervisor basic requirement when it comes to storage. This can be due to the government of regulatory directly onto the hardware. All operating systems in full virtualization compliance. EBS offers an AWS managed encryption feature. •Users can enable encryption when communicate directly with the VM hypervisor, so guest operating systems do creating EBS volumes by clicking on a checkbox. •Encryption Keys are managed by the Key not require any modification. Guest operating systems in full virtualization Management Service (KMS) provided by AWS. •Encrypted volumes can only be attached to systems are generally faster than other virtualization schemes. selected instance types. •Encryption uses the AES-256 algorithm. •Snapshots from encrypted Load balancing: One characteristic of cloud computing is virtualized network volumes are encrypted and similarly, volumes created from snapshots are encrypted. 4. EBS access to a service. No matter where the business is located, to access the Volumes are independent of the EC2 instance they are attached to. The data in an EBS volume will service, it can be directed to the available resources. The technology used to remain unchanged even if the instance is rebooted or terminated. distribute service requests to resources is referred to as load balancing. Load balancing can be implemented in hardware, through servers, or in software, such as the Squid proxy and cache daemon. 1.Load balancing is an Google Analytics is a statistical tool that measures the no. and types of visitors to a website optimization technique; it can be used to increase utilization and throughput, and how the website is used. It is offered as a free service and has been adopted by many lower latency, reduce response time, and avoid system overload. 2.Without sites. Analytics works by using JavaScript snippet called the Google Analytics Tracking load balancing, cloud computing would very difficult to manage.3.Load Code(GATC) on individual pages to implement a page tag. When the page loads, the JavaScript balancing provides the necessary redundancy to make an intrinsically runs and creates a first-party browser cookie that can be used to manage return visitors, unreliable system reliable through managed redirection.4.It also provides perform tracking test browser characteristics and request tracking code that identifies the fault tolerance when coupled with a failover mechanism. VMware vMotion location of the visitor GATC requests and stores information from the user’s account. The code enables the live migration of running virtual machines from one physical stored on the user’s system acts like a beacon and collects visitor data that it sends back to GA server to another with zero downtime, continuous service availability and servers for processing. Among the visitors that can be tracked are those that land from search complete transaction integrity. VMotion is a key enabling technology for engines, referral links in e-mail, documents, and web pages, display ads, PPC networks and creating the dynamic, automated and self optimizing datacentre. VMotion some other sources. GA aggregates the data and presents the information in a visual form. GA also is connected to the AdWords systems so that it can track the performance of particular allows users to: 1. Perform hardware maintenance without scheduled ads in different contexts. downtime. 2. Proactively migrate virtual machines away from failing or Google App Engine (GAE) is a Platform as a Service (PaaS) cloud-based Web hosting service on underperforming servers. 3.Automatically optimize and allocate entire pools Google's infrastructure. This service allows developers to build and deploy Web applications of resources for optimal hardware utilization and Alignment with business and have Google manage all the infrastructure needs, such as monitoring, failover, clustering, priorities. VMware DRS dynamically balances computing capacity across a machine instance management, and so forth. For an application to run on GAE, it must comply collection of hardware resources aggregated into logical resource pools, with Google's platform standards, which narrows the range of applications that can be run and continuously monitoring utilization across resource pools and intelligently severely limits those applications' portability. Applications running in GAE are isolated from allocating available resources among virtual machines based on pre-defined the underlying operating system, which Google describes as running in a sandbox. This allows rules that reflect businesses needs and changing priorities. When a virtual GAE to optimize the system so Web requests can be matched to the current traffic load. It also machine experiences an increased load, VMware DRS automatically allocates allows applications to be more secure because applications can connect only to computers additional resources by redistributing virtual machines among the physical using the specified URLs for the e-mail and fetch services using HTTP or HTTPS over the servers in the resource pool. VMware DRS allows IT organizations to: standard well-known ports. URL fetch uses the same infrastructure that retrieves. Web pages 1.Prioritize resources to the highest value applications in order to align on Google. The mail service also supports Gmail's messaging system. Applications also are resources with business goals. 2.Optimize hardware utilization automatically limited in that they can only read files; they cannot write to the file system directly. To access and continuously to respond to changing conditions. 3.Provide dedicated data, an application must use data stored in the memcache (memory cache) , the datastore, resources to business units while still profiting from higher hardware or some other persistent service. Memcache is a fast in- memory key-value cache that can be utilization through resource pooling. 4.Conduct zero-downtime server used between application instances. For persistent data storage of transactional data, the maintenance. VMware vSphere Distributed Switch (VDS) provides a datastore is used. Additionally, an application responds only to a specific HTTP request-in real- centralized interface from which users can configure, monitor and administer me, part of a queue, or scheduled and any request is terminated if the response requires more virtual machine access switching for the entire data center. The VDS provides: than 30 seconds to complete. GAE has a distributed datastore system that supports queries 1.Simplified virtual machine network configuration. 2.Enhanced network and transactions. This datastore is no-relational or “schema-less”, but it does store data monitoring and troubleshooting capabilities. 3.Support for advanced VMware objects or entries that are assigned properties. vSphere networking features. AdWords is a targeted ad service based on matching advertisers and their keywords to Cloud Computing is a technology that allows users to access and use shared data and users and their search profiles. This service transformed Google from a competent search computing services via the Internet or a Virtual Private Network using a scalable range engine into an industry giant and is responsible for the majority of Google's revenue of resources without having to build infrastructure to support these resources within stream. AdWords' two largest competitors are Microsoft adcenter and Yahoo! Search their own environments or networks. It uses special hardware and software to deliver Marketing. Ads are displayed as text, banners, or media an can be tailored based on a service over a network (typically the Internet). With cloud computing, users can geographical location, frequency, IP addresses, and other factors AdWords ads can access files and use applications from any device that can access the Internet. An appear not only on Google.com, but on AOL search, Ask.com, an Netscape, along with example of a Cloud Computing provider is Google's Gmail. Gmail users can access files other partners. Other partners belonging to the Google Display Network can also display and applications hosted by Google via the internet from any device. AdSense ads. In all these cases, the AdWords system determines which ads to match to Essential Characteristics: •On-demand self-service: A consumer can unilaterally provision the user searches. •Advertisers bid on keywords that are used to match a user to their computing capabilities, such as server time and network storage, as needed automatically product or service. •If a user searches for a term such as "develop abdominal muscles", without requiring human interaction with each service's provider. •Broad network access: Google return products based on those terms. •Up to 12 ads per search can be returned. Capabilities are available over the network and accessed through standard mechanisms •Google gets paid for the ad whenever a user clicks it. The system is referred to a pay- that promote use by heterogeneous thin or thick client platforms (e.g., mobile phones, per-click advertising, and the success of the ad is measured by what is called the click- laptops, and PDAs). •Resource pooling: The provider's computing resources are pooled to through rate (CTR). •Google calculates a quality score for ads based on the CTR, the strength serve multiple consumers using a multi-tenant model, with different physical and virtual of the connection between the ad and the keywords, and the advertiser's history with Google. resources dynamically assigned and reassigned according to consumer demand. There is a •This quality score is a Google trade secret and is used to price the minimum bid of keyword. sense of location-independence in that the customer generally has no control or knowledge Google Cloud: Googleis the most visited site on the Internet, with massive infrastructure over the exact location of the provided resources but may be able to specify location at a and one of the largest clouds. It operates over a million servers globally, processes a higher level of abstraction (e.g., country, state, or data center). Examples of resources billion searches, and generates 20 petabytes of data daily. Google uses hundreds of IP include storage, processing, memory, network bandwidth, and virtual machines. •Rapid addresses pooled and load-balanced in layers. Queries are sent to Google's DNS servers, elasticity: Capabilities can be rapidly and elastically provisioned, in some cases which assign the closest IP based on geography using a round-robin policy. These IPs automatically, to quickly scale out and rapidly released to quickly scale in. To the consumer, direct requests to nearby data centers and clusters containing thousands of servers the capabilities available for provisioning often appear to be unlimited and can be running a custom Linux OS. Load balancing occurs at multiple levels: 1.DNS Layer: IPs purchased in any quantity at any time. •Measured service: Cloud systems automatically assigned geographically to the nearest datacenter. 2.Cluster Level: Requests forwarded control and optimize resource use by leveraging a metering capability at some level of to Squid proxy servers and caches based on server load. 3.Server Level: Queries sent to abstraction appropriate to the type of service (e.g., storage, processing, bandwidth, and Google Web Servers based on utilization.Google Web Servers query the index and return active user accounts). Resource usage can be managed, controlled, and reported providing transparency for both the provider and consumer of the utilized service. results as HTML. The system is designed for high performance, redundancy, and failover. The Dark Web is the World Wide Web content that exists on darknets, overlay Workload management shifts tasks from failed to redundant servers, with data stored networks which use the Internet but require specific software, configurations or on redundant systems and applications running on multiple hosts. Google avoids authorization to access. Users of the dark web refer to the regular web as Clearnet hardware virtualization, relying instead on software-based load balancing. due to its unencrypted nature. The Dark Web actually refers to a set of accessible, albeit anonymously hosted, website that exist within the Deep Web. Because these Windows AZURE service: Microsoft calls their cloud operating system the Windows websites are not indexed by normal search engines, users can access them only with Azure Platform. Azure is Microsoft's Infrastructure as a Service (laaS) Web hosting special software that disguises his IP address The most common software used to service. Windows Azure serves as a cloud operating system with a fabric infrastructure access the Dark Web is The Onion Browser, referred to as TOR. The Dark Web is of virtual machines hosted in Microsoft datacenters. With Azure's architecture an much smaller than the Deep Web, and it's made up of numerous types of sites. But application can run locally, run in the cloud, or some combination of both. Applications it's perhaps most popular for its anonymous marketplace that often sell illegal on Azure can be run as applications, as background processes or services, or as both. products like drugs or weapons. The Azure Windows Services Platform API uses the industry standard REST, HTTP, and XML protocols that are part of any Service Oriented Architecture cloud infrastructure to allow applications to talk to Azure. The Azure Service Platform hosts runtime versions of .NET Framework applications written in any of the languages in common use, such as Visual Basic, C++, C#, Java, and any application that has been compiled for .NET's Common Language Runtime (CLR). Azure also can deploy Web-based applications built with ASP.NET, the Windows Communication Foundation (WCF), and PHP, and it supports Microsoft's automated deployment technologies. Simple Cloud API: Interoperability is a great concern in cloud computing. If an application is built on a platform such as Microsoft Azure, porting that application to Amazon Web Services or GoogleApps may be difficult, if not impossible. In an effort to create an interoperability standard, Zend Technologies has started an open source initiative to create a common application program interface that will allow applications to be portable. The initiative is called the Simple API for Cloud Application Services. It has its goal a set of common interfaces for: •File Storage Services: Currently Amazon S3, Windows Azure, Blob Storage, Nirvanix and Local storage is supported. •Document Storage Services: Amazon SimpleDB and Windows Azure Table Storage are currently supported. •Simple Queue Services: Amazon SQS, Windows Azure Queue Storage, and Local queue services are supported. Life Cycle of Cloud Computing: Cloud services have a defined lifecycle that a management program needs to know and implement at each of the six different stages in that lifecycle: 1. Phase 1: Define the services to be provided in a template: The tasks include creating, updating, deleting the templates that are used for creating instances in the cloud. 2. Phase 2: Interactions of the services through an SLA: The main task of this phase is to manage and maintain the clients relationships, create and manage the service contracts. 3. Phase 3: Deployment of the instance to the cloud and managing them at the runtime: This phase deals with creating, updating and deleting the service Identity management is a primary mechanism for controlling access to data in the offerings. 4. Phase 4: Defining the attributes of the operating services, and modifying cloud, preventing unauthorized uses, maintaining user roles, and complying with the properties: Few tasks like managing the operating services and modification of the regulated to Identities also are important from a security standpoint because they same are involved in this phase. 5. Phase 5: Managing the instances and performing can be authenticate client requests for services in a distributed network system such routine maintenance: The business must monitor the resources, track, respond to as the Internet or, in this case, for cloud computing services. Protocols: •OpenID 2.0 events, perform reporting and billing, and other tasks. 6. Phase 6: Retirement of the is the standard associated with creating an identity and having a third-party service service: At this phase the tasks are to protect the data, migration of the system if authenticate the use of that digital identity It is the key to creating Single Sign-On required, archiving, service contract renewal etc. (SSO) systems. Some cloud service providers have adopted OpenID as a service, and Content Management Systems (CMS): Mary websites are based on the notion of its use is growing. OpenID doesn't specify the means for authentication of an information management and organization; they are referred to as content identity, and it is up to the particular system how the authentication process is management systems (CMS). A database is a content management system, but the executed. Authentication can be by a Challenge and Response Protocol (CHAP), notion of a website as a CMS adds a number of special features to the concept that through a physical smart card, or using a flying finger or evil eye through a biometric includes rich user interaction, multiple data sources, and extensive customization and measurement. •The second protocol is used to present identity-based claims in extensibility. The Drupal CMS was chosen as an example of this type of PaaS because it cloud computing is a authorization markup languages that create files in the form of is so extensively used and has broad industry impact, and it is full-strength developer being XACML and SAML. SAML (Security Assertion Markup Language) is gaining tool. Whereas Drupal is used in major Web sites and organizes vast amounts of growing, acceptance among cloud service providers. It is a standard of OASIS and an information, the site Squarespace.com was chosen to illustrate a point- and-click CMS XML standard for passing authentication and authorization between an identity system aimed at supporting individuals, small businesses, and other small provider and the service provider. SAML is a complimentary mechanism to OpenID organizations. Squarespace is often associated with blogging tools (as is Drupal), but it and in used to create SSO systems. •An open standard called OAuth provides a is more than that. Squarespace works with photos, imports information from other token service that can be used to present validated access to resources. OAuth is social pols, and allows very attractive websites to be created by average users. similar to OpenID, but provides a different mechanism for shared access. The use of Advantages: •High reliability and scalability •Easy to use •Highly secured •Provides OAuth tokens allows clients to present credentials that contain no account flexibility to users and service providers •Easily accessible. (PTO for ERP) information (userID or password) to a cloud service. The token comes with a defined period after which can no longer be used. Several important cloud service providers Cloud storage is a repository of data in a cloud that may be accessed by web based have begun to make OAuth APIs available based on the OAuth 2.0 standard, most API's. Based on the need, the company can choose either block storage devices or file notably Facebook's Graph API and the Google Data API. storage devices. A block storage device exposes its storage to clients as raw storage that can be partitioned to create volumes. It is up to the operating system to create and manage the file system; from the standpoint of the storage device, data is transferred in blocks. The alternative type of storage is a file server, most often in the form of a Network Attached Storage (NAS) device. NAS exposes its storage to clients in the form of files, maintaining its own file system. Block storage devices offer faster data transfers, but impose additional overhead on clients. File-oriented storage devices are generally slower (with the exception of large file-streaming applications), but require Cloud Security Management is the practice of protecting cloud-based systems, data, less overhead from attached clients. Features: •It has a greater availability of resources. and applications from unauthorized access, threats, and vulnerabilities. It involves a •Easy maintenance is one of the key benefits of using Cloud computing. •Cloud comprehensive set of policies, processes, and technologies to ensure the computing has a Large Network Access. •It has an automatic system. •Security is one of confidentiality, integrity, and availability of cloud resources. Key Components of the major components and using cloud computing you can secure all over the networks. Cloud Security Management: ● Risk Assessment: Identifying and evaluating potential Cloud Storage Levels: 1. Block-Based Storage System: •Hard drives are block-based security risks specific to the cloud environment. ● Access Management: Implementing storage systems. Your operating system like Windows or Linux actually sees a hard disk controls to restrict access to cloud resources based on user roles and permissions. ● drive. So, it sees a drive on which you can create a volume, and then you can partition Data Security: Protecting sensitive data stored in the cloud, including encryption, data that volume and format them. •For example, If a system has 1000 GB of volume, then loss prevention (DLP), and backup and recovery. ● Network Security: Securing we can partition it into 800 GB and 200 GB for local C and local D drives respectively. network connections between cloud resources and on-premises systems, using •Remember with a block-based storage system, your computer would see a drive, and firewalls, intrusion detection systems (IDS), and virtual private networks (VPNs). ● then you can create volumes and partitions. 2. File-Based Storage System: •In this, you Patch Management: Ensuring that cloud systems and applications are updated with are actually connecting through a Network Interface Card (NIC). You are going over a the latest security patches to address vulnerabilities. ● Incident Response: Developing network, and then you can access the network-attached storage server (NAS). NAS a plan to respond to security incidents promptly and effectively. ● Compliance: devices are file-based storage systems. •This storage server is another computing Adhering to relevant industry standards and regulations, such as GDPR, HIPAA, and device that has another disk in it. It is already created a file system so that it’s already PCI DSS. Benefits: ●Reduced Risk: Protecting sensitive data and preventing formatted its partitions, and it will share its file systems over the network. Here, you unauthorized access ensures that businesses avoid legal liabilities and reputational harm caused by data breaches. ●Compliance Adherence: Meeting regulatory can actually map the drive to its network location. •In this, like the previous one, there requirements and avoiding fines also positions the organization for smoother audits is no need to partition and format the volume by the user. It’s already done in file- and stronger partnerships with compliance-focused clients. based storage systems. So, the operating system sees a file system that is mapped to a local drive letter. (PTO) Advantages: •Easy implementation: Cloud hosting allows business to retain the same Google API is a great set of developer's tools. These tools make programmers able to applications and business processes without having to deal with the backend perform operations using google API and living within their development framework. technicalities. Readily manageable by the Internet, a cloud infrastructure can be Google API provides a way to use its features easily. You can use its maps api, google accessed by enterprises easily and quickly. •Accessibility: Access your data anywhere, feeds api, google search api and google friends connect which is very useful tools for any anytime. An Internet cloud infrastructure maximizes enterprise productivity and site. Two programming language environments that are compatible with Google API are efficiency by ensuring your application is always accessible. This allows for easy Java and NET. •Ads and AdSense: These APIs allow Google's advertising services to be collaboration and sharing among users in multiple locations. •No hardware required: integrated into Web applications. The most commonly used services in this category are Since everything will be hosted in the cloud, a physical storage center is no longer AdWords, AdSense, and Google Analytics. •AJAX: The Google AJAX APIs provide a means needed. However, a backup could be worth looking into in the event of a disaster that to add content such as RSS feeds, maps, search boxes, and other information sources by could leave your company's productivity stagnant. •Cost per head: Overhead including a snippet of JavaScript into your code. •Browser: Google has several APIs technology costs are kept at a minimum with cloud hosting services, enabling related to building browser-based applications, including four for the Chrome browser. businesses to use the extra time and resources for improving the company This category includes the Google Cloud Print API, the Installable Web Apps API for infrastructure. •Flexibility for growth: The cloud is easily scalable so companies can creating installation packages, the Google Web Toolkit for building AJAX applications add or subtract resources based on their needs. As companies grow, their system will using Java, and V8, which is a high-performance JavaScript engine. •Data: The Data APIs grow with them. •Efficient recovery: Cloud computing delivers faster and more are those that exchange data with a variety of Google services. The list of Google Data accurate retrievals of applications and data. With less downtime, it is the most APIs includes Google Apps, Google Analytics, Blogger, Base, Book, Calendar, Code Search, efficient recovery plan. Google Earth, Google Spreadsheets, Google Notebook, and Picasa Web Albums. •Geo: A Disadvantages: •No longer in control: When moving services to the cloud, the number of APIs exist to give location-specific information hooking into maps and geo- company is handing over the data and information. For companies who have an in- specific databases. Some of the more popular APIs in this category include Google Earth, house IT staff, they will be unable to handle issues on their own. •Security: Although Directions, JavaScripts Maps, Maps API for Flash, and Static Maps. •Search: The search cloud service providers implement the best security standards and industry APIs leverage Google's core competency and its central service. APIs such as Google AJAX certifications, storing data and important files on external service providers always Search, Book Search, Code Search, Custom Search, and Webmaster Tools Data APIs allow opens up risks. Using cloud-powered technologies means one will need to provide developers to include Google searches in their applications and web sites. •Social: Many service provider with access to important business data. Meanwhile. being a public Google APIs are used for information exchange and communication tools. They support service opens up cloud service providers to security challenges on a routine basis. The applications such as Gmail, Calendar, and others, and they provide a set of foundation ease in procuring and accessing cloud services can also give nefarious users the ability services. The popular social APIs are Blogger Data, Calendar, Contacts, OpenSocial, to scan, identify and exploit loopholes and vulnerabilities within a system. Picasa, and YouTube. •Dependency and vendor lock-in: One of the major disadvantages of cloud computing is the implicit dependency on the provider. This is what the industry calls "vendor lock- in” since it is difficult, and sometimes impossible, to migrate from a provider once the company has rolled with him. If a user wishes to switch to some other provider, then Azure SQL Database is a relational database (RDBMS) service provided by Microsoft can be really painful and cumbersome to transfer huge data from the old provider to Azure that is widely used by developers when creating new applications in the cloud. It the new one. •No Redundancy: A cloud server is not redundant nor is it backed up. As is managed completely by Microsoft and is a highly scalable platform-as-a-service technology may fail here and there one must try to avoid getting burned by (PaaS) designed especially for cloud applications. Users can create databases, organize purchasing a redundancy plan. Although it is an extra cost, in most cases it will be well data into tables, and run queries. Here, we create a managed database server in the worth it. •Bandwidth issues: For ideal performance, clients have to plan accordingly and cloud and use the server to deploy our database. The server is a logical construct that not pack large amounts of servers and storage devices into a small set of data centers. acts as the central administration point for pooled databases or multiple logins, auditing rules, threat detection policy, and failover groups. The databases are available Cloud Security Alliance(CSA): Cloud Security Model: Service in cloud computing as Single databases and elastic pools: •Single Database: This option helps the includes laaS, PaaS, SaaS etc. The security of the services vary to a great extent, from developers instantly get started with a single SQL Server database by creating and each other. For example, laas has the lowest level of security, whereas the other two running it in the cloud and accessing this database through the server. It is is more secured. As one moves upward in the stack, each service model inherits the a PaaS offering so everything is managed by Microsoft, so all we have to do is to capabilities of the model beneath it, as well as all the inherent security concerns and configure the database, create the necessary tables required to carry out the risk factors. laas supplies the infrastructure; PaaS adds application development operations, and fill in the required data. We can scale the database as per our frameworks, transactions, and control structures; and SaaS is an operating requirements (if we need more storage, memory, and processing power). By default, environment with applications, management, and the user interface. As one ascends the stack, laas has the least levels of integrated functionality and the lowest levels of Microsoft pre-allocates some resources and we are charged per hour based on the integrated security, and Saas has the most. The most important lesson from this resources we are using. •Elastic Pool: It is similar to single databases that we have discussion of architecture is that each different type of cloud service delivery model talked about above, except that by default multiple databases can share the same creates a security boundary at which the cloud service provider's responsibilities end resources (memory, storage space, processing power) through multiple tenancies. Here and the customer's responsibilities begin. Any security mechanism below the security the different resources are referred to as a pool. This model is very useful when we boundary must be built into the system, and any security mechanism above must be have databases with resource requirements varying with time as it helps allocate and maintained by the customer. As one move up the stack, it becomes more important deallocate the resources as per our needs thus reducing costs and helping us be quick to make sure that the type and level of security is part of your Service Level and efficient. It enables us to use resources available in the created pool and then Agreement. release them once processing is complete. In the SaaS model, the vendor provides security as part of the Service Level agreement, with the compliance, governance and liability levels stipulated under the A Network Management System (NMS) is a platform that enables the monitoring and contract for the entire stack. For the PaaS model, the security boundary may be management of a network. A NMS lets network engineers manage a network’s defined for the vendor to include the software framework and middleware layer. In independent components inside a bigger network management framework and the PaaS model, the customer would be responsible for the security of the application performs several key functions. A NMS identifies, configures, monitors, updates and and UI at the top of the stack. The model with the least built-in security is IaaS, where troubleshoots network devices — both wired and wireless — in an enterprise network. everything that involves software of any kind is the customer’s problem. A system management control application then displays the performance data A Service Oriented Architecture (SOA) describes a standard method for requesting collected from each network component, allowing network engineers to make changes services from distributed components and managing the results. The clients as needed. Eg, BMC PATROL, CA Unicenter, IBM Tivoli, HP Open View, and Microsoft requesting services, the components providing the services, the protocols used to System Center. NMS functions: •Performance monitoring. By collecting operating deliver messages, and the responses can vary widely, hence SOA is required. SOA is a metrics through a series of physical taps, software agents or Simple Network specification and a methodology for providing platform- and language-independent Management Protocol interfaces, NMSes provide visibility to determine if network services for use in distributed applications. SOA provides the translation and elements operate correctly. •Device detection. NMSes detect devices on the network management layer in an architecture that removes the barrier for a client obtaining and ensures the network recognizes devices and configures them correctly. desired services. With SOA, clients and components can be written in different •Performance analysis. NMSes track performance data indicators, including bandwidth languages and can use multiple messaging protocols and networking protocols to utilization, packet loss, latency, availability and uptime of network components. communicate with one another. SOA provides the standards that transport the •Notification alerts: In the event of a system disruption, NMSes proactively alert messages and provides access to reusable Web services over a TCP/IP network, as administrators about any performance issues. Features: Administration of resources, well. Eg.SOAP, REST. An ESB is a architectural pattern comprised of a set of network Configuring resources, Enforcing security, Monitoring operations, Optimizing services that manage transactions in a Service Oriented Architecture. In the figure, performance, Policy management, Performing maintenance, Provisioning of resources. three hypothetical different applications are shown interfaced with an authentication Enterprise Resource Planning (ERP) and is a software similar to CRM that is hosted on module through an Enterprise Service Bus. Messages flow from client to component cloud servers which helps the enterprises to manage and manipulate their business through the ESB, which manages these transactions, even though the location of the data as per their needs and user requirements. ERP software follows pay per use services comprising the ESB may vary widely. An ESB therefore plays the role of a methodologies of payment, that is at the end of the month, the enterprise pay the transaction broker in SOA. ensuring that messages go where they are supposed to go amount as per the cloud resources utilized by them. There are various ERP vendors and are acted upon properly. The service bus performs the function of mediation: available like Oracle, SAP, Epicor, SAGE, Microsoft Dynamics, Lawson Softwares and message translation, registration, routing, logging, auditing, and managing many more. Advantages: •Cost effective •High mobility •Increase in productivity •No transactional integrity. Transactional integrity is similar to ACID in a database system security issues •Scalable and efficient. Salesforce is a cloud-based software atomicity, consistency, isolation, and durability the essence of which is that platform designed to help businesses find prospects, close deals, and deliver transactions succeed or they fail and are rolled hack. An ESB creates a virtual exceptional customer service. It combines CRM capabilities with robust cloud services environment layered on top of an enterprise messaging system where services are to cater to various business needs, including sales, marketing, analytics, and more. advertised and accessed. Features: •Monitoring services aid in managing events. •Process management services manage message transactions. •Data repositories or ….3. Object-Based Storage System: •In this, a user uploads objects using a web browser registries store business logic and aid in governance of business processes. •Data and uploads an object to a container i.e., Object Storage Container. This uses the HTTP services pass messages between clients and services. •Data abstraction services Protocols with the rest of the APIs (for example: GET, PUT, POST, SELECT, DELETE). •For translate messages from one format to another, as required. •Governance is a service example, when you connect to any website, you need to download some images, text, or that monitors compliance of the business operations with governmental regulation, anything that the website contains. For that, it is a code HTTP GET request. If you want which can vary from state to state and from country to country. •Security services to review any product then you can use PUT and POST requests. •Also, there is no validate clients and services and allow messages to pass from one to the other. hierarchy of objects in the container. Every file is on the same level in an Object-Based storage system. ●Enhanced Reputation: Building trust with customers and stakeholders translates Mobility patterns in cloud computing refer to the movement and accessibility of into increased customer retention, referrals, and competitive advantage. ●Improved applications, data, and services across various environments, devices, and locations. Business Continuity: Ensuring that cloud services remain available in case of security These patterns are essential for ensuring seamless user experiences and efficient incidents minimizes downtime and prevents disruptions to critical business resource utilization. •P2V (Physical-to-Virtual) moves workloads from physical servers to operations. ●Cost Savings: Avoiding costly data breaches and security incidents virtual machines, enhancing resource utilization and scalability. Tools: VMware vCenter allows organizations to allocate resources more effectively toward growth and Converter and Microsoft Virtual Machine Converter. •V2V (Virtual-to-Virtual) migrates innovation initiatives. AppZero Virtual Application Appliance: Applications that run workloads between virtual environments or clouds, supporting cross-cloud strategies in datacenters are captive to the operating systems and hardware platforms that they and upgrades. Tools: VMware vMotion and CloudEndure. •V2P (Virtual-to-Physical) run on. So moving an application from one platform to another nearly as simple as shifts workloads from virtual systems to physical hardware for performance-critical moving a machine image from one system to another. The situation is further needs. Tools: Custom migration scripts or vendor-specific tools. •P2P (Physical-to- complicated by the fact applications are tightly coupled with the operating systems Physical) transfers workloads between physical servers for upgrades or failovers. Tools: on which they run. An application running on Windows, for example, isn’t isolated Hardware cloning tools and network-based replication software. •D2C (Datacenter-to- from other applications. When the application loads, it often loads or uses different Cloud) migrates data or applications from on-premises datacenters to cloud Dynamic Link Libraries (DLL), and it is through the sharing or modification of DLLs that environments, enabling scalability and flexibility. Tools: AWS Snowball, Azure Migrate, Windows applications get themselves in trouble. Further modifications include and Google Transfer Appliance. •C2C (Cloud-to-Cloud) transitions workloads across cloud modifying the registry during installation. These factors make it difficult to port providers or regions for multi-cloud strategies and cost optimization. Tools: CloudEndure applications from one platform to another without lots of careful work. The ability to and Multicloud Management Platforms. •C2D (Cloud-to-Datacenter) moves workloads run an application from whatever platform you want is one of the characteristics of or data from cloud environments back to on-premises datacenters, often for compliance cloud computing. While the Simple Cloud API is useful for applications written in PHP, or latency-sensitive needs. Tools: Hybrid cloud platforms like VMware Cloud on AWS or other methods may be needed to make applications easily portable. One company Azure Arc. •D2D (Datacenter-to-Datacenter) transfers workloads or data between on- working on is problem is AppZero and its solution is called the Virtual Application premises datacenters for disaster recovery, failover, or hardware upgrades. Tools: SAN Appliance (VAA). replication, Veeam Backup & Replication, or Zerto Virtual Replication.