Puppet is IT automation tool. I was a speaker for this presentation in one of Meetup and it was received well. Sharing it with open source folks who want to collaborate, learn and Win.
Docker Engine 1.12 can be rightly called ” A Next Generation Docker Clustering & Distributed System”. Though Docker Engine 1.12 Final Release is around corner but the recent RC3 brings lots of improvements and exciting features. One of the major highlight of this release is Docker Swarm Mode which provides powerful yet optional ability to create coordinated groups of decentralized Docker Engines. Swarm Mode combines your engine in swarms of any scale. It’s self-organizing and self-healing. It enables infrastructure-agnostic topology.The newer version democratizes orchestration with out-of-box capabilities for multi-container on multi-host app deployments.
Docker Online Meetup #28: Production-Ready Docker SwarmDocker, Inc.
presented by Alexandre Beslic (@abronan)
Swarm v1.0 is now ready for running your apps in production!
Swarm is the easiest way to run Docker applications at large scale on a cluster. It turns a pool of Docker Engines into a single, virtual Engine. You don’t have to worry about where to put containers, or how they’re going to talk to each other - it just handles all that for you.
We’ve spent the last few months tirelessly hardening and tuning it, and in combination with multi-host networking and the new volume system in Docker Engine 1.9, we can confidently say that it’s ready for running your apps in production. In our tests, we’ve been running Swarm on EC2 with 1,000 nodes and 30,000 containers and it keeps on scheduling containers in less than half a second. Not even breaking a sweat! Keep an eye for a blog post soon with the full details.
Read more: http://blog.docker.com/2015/11/swarm-1-0/
Orchestrating Linux Containers while tolerating failuresDocker, Inc.
lthough containers are bringing a refreshing flexibility when deploying services in production, the management of those containers in such an environment still requires special care in order to keep the application up and running. In this regard, orchestration platforms like Docker, Kubernetes and Nomad have been trying to alleviate this responsibility, facilitating the task of deploying and maintaining the entire application stack in its desired state. This ensures that a service will be always running, tolerating machine failures, network erratic behavior or software updates and downtime. The purpose of this talk is to explain the mechanisms and architecture of the Docker Engine orchestration platform (using a framework called swarmkit) to tolerate failures of services and machines, from cluster state replication and leader-election to container re-scheduling logic when a host goes down.
Swarm allows multiple Docker hosts to be clustered together into a single virtual Docker host. The document discusses how to set up a Swarm cluster in three steps: 1) Create a key-value store, 2) Run the Swarm manager container, and 3) Configure Docker daemon arguments on hosts. It also provides an example of running a microservices voting application on Swarm and references additional example repositories.
The document is a slide deck presentation by Bret Fisher on going into production with Docker and Swarm. Some key points from the presentation include focusing first on Dockerfiles rather than complex orchestration, avoiding anti-patterns like using the "latest" tag or trapping unique data in containers, and starting with a simple 3 node Swarm cluster for high availability before scaling up further. The presentation also provides examples of full tech stacks using various open source and commercial tools for a Dockerized infrastructure.
Docker Swarm Is Dead: Long Live Docker SwarmElton Stoneman
From the Docker London MeetUp, presented on 27th June 2016. A walkthrough of Swarm Mode in Docker 1.12, the presentation introduces demos for creating a Docker Swarm using Azure virtual machines, and running a distributed application with a Node REST API, feeding analytics into Elasticsearch via a Redis queue.
In this overview presented to a gathering of directors for a large network equipment manufacturer, Chris discusses Docker, DevOps workflows, considerations for containers in production, and the extended Docker technology ecosystem.
Load Balancing Apps in Docker Swarm with NGINXNGINX, Inc.
On-demand webinar recording: http://bit.ly/2mRjk2g
Docker and other container technologies continue to gain in popularity. We recently surveyed the broad community of NGINX and NGINX Plus users and found that two-thirds of organizations are either investigating containers, using them in development, or using them in production. Why? Because abstracting your applications from the underlying infrastructure makes developing, distributing, and running software simpler, faster, and more robust than ever before.
But when you move from running your app in a development environment to deploying containers in production, you face new challenges – such as how to effectively run and scale an application across multiple hosts with the performance and uptime that your customers demand.
The latest Docker release, 1.12, supports multihost container orchestration, which simplifies deployment and management of containers across a cluster of Docker hosts. In a complex environment like this, load balancing plays an essential part in delivering your container-based application with reliability and high performance.
Join us in this webinar to learn:
* The basic built-in load balancing options available in Docker Swarm Mode
* The pros and cons of moving to an advanced load balancer like NGINX
* How to integrate NGINX and NGINX Plus with Swarm Mode to provide an advanced load-balancing solution for a cluster with orchestration
* How to scale your Docker-based application with Swarm Mode and NGINX Plus
Since its first 1.12 release on July 2016, Docker Swarm Mode has matured enough as a clustering and scheduling tool for IT administrators and developers who can easily establish and manage a cluster of Docker nodes as a single virtual system. Swarm mode integrates the orchestration capabilities of Docker Swarm into Docker Engine itself and help administrators and developers with the ability to add or subtract container iterations as computing demands change. With sophisticated but easy to implement features like built-in Service Discovery, Routing Mesh, Secrets, declarative service model, scaling of the services, desired state reconciliation, scheduling, filters, multi-host networking model, Load-Balancing, rolling updates etc. Docker 17.06 is all set for production-ready product today. Join me webinar organised by Docker Izmir, to get familiar with the current Swarm Mode capabilities & functionalities across the heterogeneous environments.
Docker Birthday #3 - Intro to Docker SlidesDocker, Inc.
High level overview of Docker + Birthday #3 overview (app and challenge portion)!
Learn more about Docker Birthday #3 celebrations here: https://www.docker.com/community/docker-birthday-3
The document discusses Docker 1.5 and its new features including IPv6 support, read-only containers, Docker stats, and the Docker image specification. IPv6 can be enabled by running Docker with the --ipv6 flag. Read-only containers restrict writes to containers. Docker stats provides live metrics for containers. The Docker image specification defines the format for layered image files and metadata.
It is the understanding of docker and how it was useful in day-to-day tasks, some basic difference between old and new. At last my learnings and mistake while doing it all.
Docker Datacenter Overview and Production Setup SlidesDocker, Inc.
An overview on Docker Data Center and Universal Control Plane. We will cover how to install for production and integrate Docker Trusted Registry.
Led by DDC + UCP Champ:
Vivek Saraswat
Experience Level: Attendees need no prior experience with Docker, but should be familiar with basic linux command-line.
Swarm allows multiple Docker hosts to be clustered together into a single virtual Docker host. It provides features like scheduling, rescheduling on failure, high availability with multiple masters, and service discovery. To set up a Swarm cluster, run the Swarm manager container on one host and restart the Docker daemons on the other hosts with arguments to join the cluster. An example voting app microservices demo shows how to deploy an application across a Swarm cluster.
Fluentd is an open source data collector that allows for unified logging and data collection. It can be used to collect and parse logs from multiple sources like applications and servers running on multiple hosts. Fluentd works with Docker to provide a logging driver that routes container output to Fluentd. This allows Fluentd to collect logs from Docker containers and structure the data as JSON. Fluentd then reliably forwards the logs to a destination like Elasticsearch for storage and analysis. The document demonstrates how to set up Fluentd and Elasticsearch Docker containers to collect logs produced by other application containers running on the host.
Heart of the SwarmKit: Store, Topology & Object ModelDocker, Inc.
Heart of the SwarmKit: Store, Topology & Object Model by Aaron, Andrea, Stephen D (Docker)
Swarmkit repo - https://github.com/docker/swarmkit
Liveblogging: http://canopy.mirage.io/Liveblog/SwarmKitDDS2016
Docker Online Meetup: Infrakit update and Q&ADocker, Inc.
While working on Docker for AWS and Azure, we realized the need for a standard way to create and manage infrastructure state that was portable across any type of infrastructure, from different cloud providers to on-prem. One challenge is that each vendor has differentiated IP invested in how they handle certain aspects of their cloud infrastructure. It is not enough to just provision five servers; what IT ops teams need is a simple and consistent way to declare the number of servers, what size they should be, and what sort of base software configuration is required. And in the case of server failures (especially unplanned), that sudden change needs to be reconciled against the desired state to ensure that any required servers are re-provisioned with the necessary configuration. We started InfraKit to solves these problems and to provide the ability to create a self healing infrastructure for distributed systems.
Cloud Foundry is an open source platform as a service (PaaS) that supports building, deploying, and running applications on the cloud. It supports multiple frameworks like Java, Ruby, Scala, and Node.js and services like SQL, NoSQL, messaging, and analytics. Cloud Foundry uses a distributed architecture with no single point of failure and provides automatic scaling and self-healing capabilities.
The document discusses Docker Swarm, a Docker container orchestration tool. It provides an overview of key Swarm features like cluster management, service discovery, load balancing, rolling updates and high availability. It also discusses how to deploy applications using Swarm, including accessing GPUs, the deployment workflow, and using Swarm on ARM architectures. The conclusion states that the best orchestration tool depends on one's use case and preferences as each has advantages and disadvantages.
Catching up with what has happened with logging in Docker since late 2014 all the way up to the recently released Docker 0.10. Also, presenting my view on a comprehensive approach to monitoring Docker using the API to get events, logs, stats with a little bit of self promotion in pointing out that we have recently released an implementation of comprehensive monitoring as part of a Sumo Logic collector source.
What's New in Docker 1.12 (June 20, 2016) by Mike Goelzer & Andrea LuzzardiMike Goelzer
Docker 1.12 introduces several new features for managing containerized applications at scale including Docker Swarm mode for native clustering and orchestration. Key features include services that allow defining and updating distributed applications, a built-in routing mesh for load balancing between nodes, and security improvements like cryptographic node identities and TLS encryption by default. The document also discusses plugins, health checks, and distributed application bundles for declaring stacks of services.
Using Docker Swarm Mode to Deploy Service Without Loss by Dongluo Chen & Nish...Docker, Inc.
Talk from Docker SF Meetup #50
Abstract:
Docker swarm mode enables users to manage their applications with service primitives. In this talk we demonstrate how to do service upgrades without impacting your application. The Healthcheck feature provides health indication for a container. Coming up in Docker 1.13 release, Docker Swarm can connect healthcheck result with load balancer to implement no-loss service upgrade.
Speaker Biographies:
Nishant Totla is a software engineer at Docker, and works on the core open source team. He is currently working on Docker SwarmKit and Docker Swarm. Prior to Docker, he was a PhD student at UC Berkeley, doing research on programming languages. In his spare time, he enjoys long-distance running, biking, and other outdoor activities. Nishant tweets at @nishanttotla.
Dongluo Chen is a software engineer at Docker focusing on orchestration and container development. Before Docker he was software engineer manager at Microsoft Azure building and automating global data centers. He worked at France Telecom (Orange) and the Ohio State University as research scientist in networking area.
- The document introduces Docker, explaining that it provides standardization for packaging software applications and dependencies to make them portable and help transition from monolithic to microservices architectures.
- Docker uses operating system-level virtualization rather than hardware virtualization, allowing containers to start quickly while remaining isolated and using fewer resources than virtual machines.
- The document covers Docker concepts like images, containers, and registries, and provides examples of common Docker commands for pulling images, running containers, and managing the Docker workflow of building, shipping, and running applications.
This document provides an overview of Docker Swarm and how to set up and use a Docker Swarm cluster. It discusses key Swarm concepts, initializing a cluster, adding nodes, deploying services, rolling updates, draining nodes, failure scenarios, and the Raft consensus algorithm used for leader election in Swarm mode. The document walks through examples of creating a Swarm, adding nodes, deploying a service, inspecting and scaling services, rolling updates, and draining nodes. It also covers failure scenarios for nodes and managers and how the Swarm handles them.
Puppet at Scale – Case Study of PayPal's Learnings - PuppetConf 2013Puppet
"Puppet at Scale – Case Study of PayPal's Learnings" by Stan Hsu, Senior Dev Manager, PayPal.
Presentation Overview: Large scale and app level management pose challenges to any implementation of puppet. Come and learn some of the challenges PayPal Deployment Systems team faced and the how these were overcome.
Speaker Bio: Stan Hsu is the Senior Dev Manager for PayPal's deployment systems team. His team is currently responsible to build out a new deployment system based on puppet. In his tenure at eBay/PayPal, he's had the unique experience of having had access to all data centers in both eBay and PayPal to help build out of new deployment systems for production and QA environments. His interests include application at scale, scalability, performance tuning, and usability. In his previous roles he has managed teams at Tibco, Crossworlds, and HP.
As a PHP developer building web applications is besides making a living a lot of fun too, especially when you can deploy your apps to any kind of environment and on any platform. In this session I take a non-standard PHP application (based on Zend Framework) and deploy it to a bare metal environment running LAMP, Windows 2008 Server with IIS7 and to cloud instances like Azure and Amazon.
The goal is to provide information on how to deploy to these various environments manual and automatic, but also to show it doesn't really matter anymore what the targeted platform is, as long the apps are written in PHP.
In this overview presented to a gathering of directors for a large network equipment manufacturer, Chris discusses Docker, DevOps workflows, considerations for containers in production, and the extended Docker technology ecosystem.
Load Balancing Apps in Docker Swarm with NGINXNGINX, Inc.
On-demand webinar recording: http://bit.ly/2mRjk2g
Docker and other container technologies continue to gain in popularity. We recently surveyed the broad community of NGINX and NGINX Plus users and found that two-thirds of organizations are either investigating containers, using them in development, or using them in production. Why? Because abstracting your applications from the underlying infrastructure makes developing, distributing, and running software simpler, faster, and more robust than ever before.
But when you move from running your app in a development environment to deploying containers in production, you face new challenges – such as how to effectively run and scale an application across multiple hosts with the performance and uptime that your customers demand.
The latest Docker release, 1.12, supports multihost container orchestration, which simplifies deployment and management of containers across a cluster of Docker hosts. In a complex environment like this, load balancing plays an essential part in delivering your container-based application with reliability and high performance.
Join us in this webinar to learn:
* The basic built-in load balancing options available in Docker Swarm Mode
* The pros and cons of moving to an advanced load balancer like NGINX
* How to integrate NGINX and NGINX Plus with Swarm Mode to provide an advanced load-balancing solution for a cluster with orchestration
* How to scale your Docker-based application with Swarm Mode and NGINX Plus
Since its first 1.12 release on July 2016, Docker Swarm Mode has matured enough as a clustering and scheduling tool for IT administrators and developers who can easily establish and manage a cluster of Docker nodes as a single virtual system. Swarm mode integrates the orchestration capabilities of Docker Swarm into Docker Engine itself and help administrators and developers with the ability to add or subtract container iterations as computing demands change. With sophisticated but easy to implement features like built-in Service Discovery, Routing Mesh, Secrets, declarative service model, scaling of the services, desired state reconciliation, scheduling, filters, multi-host networking model, Load-Balancing, rolling updates etc. Docker 17.06 is all set for production-ready product today. Join me webinar organised by Docker Izmir, to get familiar with the current Swarm Mode capabilities & functionalities across the heterogeneous environments.
Docker Birthday #3 - Intro to Docker SlidesDocker, Inc.
High level overview of Docker + Birthday #3 overview (app and challenge portion)!
Learn more about Docker Birthday #3 celebrations here: https://www.docker.com/community/docker-birthday-3
The document discusses Docker 1.5 and its new features including IPv6 support, read-only containers, Docker stats, and the Docker image specification. IPv6 can be enabled by running Docker with the --ipv6 flag. Read-only containers restrict writes to containers. Docker stats provides live metrics for containers. The Docker image specification defines the format for layered image files and metadata.
It is the understanding of docker and how it was useful in day-to-day tasks, some basic difference between old and new. At last my learnings and mistake while doing it all.
Docker Datacenter Overview and Production Setup SlidesDocker, Inc.
An overview on Docker Data Center and Universal Control Plane. We will cover how to install for production and integrate Docker Trusted Registry.
Led by DDC + UCP Champ:
Vivek Saraswat
Experience Level: Attendees need no prior experience with Docker, but should be familiar with basic linux command-line.
Swarm allows multiple Docker hosts to be clustered together into a single virtual Docker host. It provides features like scheduling, rescheduling on failure, high availability with multiple masters, and service discovery. To set up a Swarm cluster, run the Swarm manager container on one host and restart the Docker daemons on the other hosts with arguments to join the cluster. An example voting app microservices demo shows how to deploy an application across a Swarm cluster.
Fluentd is an open source data collector that allows for unified logging and data collection. It can be used to collect and parse logs from multiple sources like applications and servers running on multiple hosts. Fluentd works with Docker to provide a logging driver that routes container output to Fluentd. This allows Fluentd to collect logs from Docker containers and structure the data as JSON. Fluentd then reliably forwards the logs to a destination like Elasticsearch for storage and analysis. The document demonstrates how to set up Fluentd and Elasticsearch Docker containers to collect logs produced by other application containers running on the host.
Heart of the SwarmKit: Store, Topology & Object ModelDocker, Inc.
Heart of the SwarmKit: Store, Topology & Object Model by Aaron, Andrea, Stephen D (Docker)
Swarmkit repo - https://github.com/docker/swarmkit
Liveblogging: http://canopy.mirage.io/Liveblog/SwarmKitDDS2016
Docker Online Meetup: Infrakit update and Q&ADocker, Inc.
While working on Docker for AWS and Azure, we realized the need for a standard way to create and manage infrastructure state that was portable across any type of infrastructure, from different cloud providers to on-prem. One challenge is that each vendor has differentiated IP invested in how they handle certain aspects of their cloud infrastructure. It is not enough to just provision five servers; what IT ops teams need is a simple and consistent way to declare the number of servers, what size they should be, and what sort of base software configuration is required. And in the case of server failures (especially unplanned), that sudden change needs to be reconciled against the desired state to ensure that any required servers are re-provisioned with the necessary configuration. We started InfraKit to solves these problems and to provide the ability to create a self healing infrastructure for distributed systems.
Cloud Foundry is an open source platform as a service (PaaS) that supports building, deploying, and running applications on the cloud. It supports multiple frameworks like Java, Ruby, Scala, and Node.js and services like SQL, NoSQL, messaging, and analytics. Cloud Foundry uses a distributed architecture with no single point of failure and provides automatic scaling and self-healing capabilities.
The document discusses Docker Swarm, a Docker container orchestration tool. It provides an overview of key Swarm features like cluster management, service discovery, load balancing, rolling updates and high availability. It also discusses how to deploy applications using Swarm, including accessing GPUs, the deployment workflow, and using Swarm on ARM architectures. The conclusion states that the best orchestration tool depends on one's use case and preferences as each has advantages and disadvantages.
Catching up with what has happened with logging in Docker since late 2014 all the way up to the recently released Docker 0.10. Also, presenting my view on a comprehensive approach to monitoring Docker using the API to get events, logs, stats with a little bit of self promotion in pointing out that we have recently released an implementation of comprehensive monitoring as part of a Sumo Logic collector source.
What's New in Docker 1.12 (June 20, 2016) by Mike Goelzer & Andrea LuzzardiMike Goelzer
Docker 1.12 introduces several new features for managing containerized applications at scale including Docker Swarm mode for native clustering and orchestration. Key features include services that allow defining and updating distributed applications, a built-in routing mesh for load balancing between nodes, and security improvements like cryptographic node identities and TLS encryption by default. The document also discusses plugins, health checks, and distributed application bundles for declaring stacks of services.
Using Docker Swarm Mode to Deploy Service Without Loss by Dongluo Chen & Nish...Docker, Inc.
Talk from Docker SF Meetup #50
Abstract:
Docker swarm mode enables users to manage their applications with service primitives. In this talk we demonstrate how to do service upgrades without impacting your application. The Healthcheck feature provides health indication for a container. Coming up in Docker 1.13 release, Docker Swarm can connect healthcheck result with load balancer to implement no-loss service upgrade.
Speaker Biographies:
Nishant Totla is a software engineer at Docker, and works on the core open source team. He is currently working on Docker SwarmKit and Docker Swarm. Prior to Docker, he was a PhD student at UC Berkeley, doing research on programming languages. In his spare time, he enjoys long-distance running, biking, and other outdoor activities. Nishant tweets at @nishanttotla.
Dongluo Chen is a software engineer at Docker focusing on orchestration and container development. Before Docker he was software engineer manager at Microsoft Azure building and automating global data centers. He worked at France Telecom (Orange) and the Ohio State University as research scientist in networking area.
- The document introduces Docker, explaining that it provides standardization for packaging software applications and dependencies to make them portable and help transition from monolithic to microservices architectures.
- Docker uses operating system-level virtualization rather than hardware virtualization, allowing containers to start quickly while remaining isolated and using fewer resources than virtual machines.
- The document covers Docker concepts like images, containers, and registries, and provides examples of common Docker commands for pulling images, running containers, and managing the Docker workflow of building, shipping, and running applications.
This document provides an overview of Docker Swarm and how to set up and use a Docker Swarm cluster. It discusses key Swarm concepts, initializing a cluster, adding nodes, deploying services, rolling updates, draining nodes, failure scenarios, and the Raft consensus algorithm used for leader election in Swarm mode. The document walks through examples of creating a Swarm, adding nodes, deploying a service, inspecting and scaling services, rolling updates, and draining nodes. It also covers failure scenarios for nodes and managers and how the Swarm handles them.
Puppet at Scale – Case Study of PayPal's Learnings - PuppetConf 2013Puppet
"Puppet at Scale – Case Study of PayPal's Learnings" by Stan Hsu, Senior Dev Manager, PayPal.
Presentation Overview: Large scale and app level management pose challenges to any implementation of puppet. Come and learn some of the challenges PayPal Deployment Systems team faced and the how these were overcome.
Speaker Bio: Stan Hsu is the Senior Dev Manager for PayPal's deployment systems team. His team is currently responsible to build out a new deployment system based on puppet. In his tenure at eBay/PayPal, he's had the unique experience of having had access to all data centers in both eBay and PayPal to help build out of new deployment systems for production and QA environments. His interests include application at scale, scalability, performance tuning, and usability. In his previous roles he has managed teams at Tibco, Crossworlds, and HP.
As a PHP developer building web applications is besides making a living a lot of fun too, especially when you can deploy your apps to any kind of environment and on any platform. In this session I take a non-standard PHP application (based on Zend Framework) and deploy it to a bare metal environment running LAMP, Windows 2008 Server with IIS7 and to cloud instances like Azure and Amazon.
The goal is to provide information on how to deploy to these various environments manual and automatic, but also to show it doesn't really matter anymore what the targeted platform is, as long the apps are written in PHP.
Build your own clouds with Chef and MCollectiveJonathan Weiss
One important part of the DevOps movement is infrastructure automation, especially if you are running your application on top of services like Amazon EC2.
Everybody's dream is to be able to bootstrap and deploy hundreds or even thousands of machines with a few simple commands. This talk will tell you how you can do this using Open Source tools like Chef and mcollective. Chef manages your servers configuration using a nice Ruby DSL while mcollective orchestrates and commands all your nodes.
Jesus Nunez presented on using Puppet in a decentralized, masterless architecture. In a masterless configuration, each node pulls Puppet code and modules from a central Git repository and compiles its own catalog without a master server. A remote executor is used to trigger Puppet runs on nodes by updating code via Librarian Puppet, generating Puppet and ENC files, and running Puppet apply commands. This allows distributed processing without a single point of failure compared to a traditional master-node architecture.
Introduction to orchestration using McollectivePuppet
"Introduction to orchestration using MCollective" by Pieter Loubser at Puppet Camp London 2013. Find the video here: http://puppetlabs.com/community/puppet-camp
Continuous Deployment Applied at MyHeritageRan Levy
Learn how continuous deployment was applied at MyHeritage. Check out how we automated the process using MCollective, RPM, Jenkins, unit and integration tests using JUnit, PHPUnit, Cucumber and more.
PuppetConf track overview: Modern InfrastructurePuppet
From containers to Docker, Mesos and Kubernetes — you'll hear about it at PuppetConf 2016 in San Diego. Learn more and register at https://puppet.com/puppetconf/.
11 Ways to Hack Puppet for Fun and Productivity - Luke Kanies - Velocity 2012Puppet
Luke Kanies, founder and CEO of Puppet Labs, talks at Velocity '12 on 11 ways to hack Puppet for fun and productivity. http://www.puppetlabs.com
LAS16-301: OpenStack on Aarch64, running in production, upstream improvements...Linaro
LAS16-301: OpenStack on Aarch64, running in production, upstream improvements, and interoperability
Speakers: Yibo Cai, Gema Gomez Solano, Jack He, Marcin Juskiewicz, Martin Stadtler
Date: September 28, 2016
★ Session Description ★
“OpenStack is at the heart of the next generation of the opensource
cloud on a global scale. During this presentation, we will touch on three themes, running an OpenStack based cloud in production by Gema Gomez and Andy Doan, followed by Marcin talking about the packaging and bug fixing on archives required to make that happen on AArch64. Jack He and Yibo Cai, will explain what it is like working with the the upstream project, the development environment, the current patches and what needs to be done next. Then Gema Gomez will Introduce the OpenStack Interop Working Group. Why is interoperability important for OpenStack? And What is Linaro doing to improve the interoperability of OpenStack?
★ Resources ★
Etherpad: pad.linaro.org/p/las16-301
Presentations & Videos: http://connect.linaro.org/resource/las16/las16-301/
★ Event Details ★
Linaro Connect Las Vegas 2016 – #LAS16
September 26-30, 2016
http://www.linaro.org
http://connect.linaro.org
Build cloud native solution using open source Nitesh Jadhav
Build cloud native solution using open source. I have tried to give a high level overview on How to build Cloud Native using CNCF graduated software's which are tested, proven and having many reference case studies and partner support for deployment
Kubernetes relies on API calls and standard Linux tools for troubleshooting. Monitoring metrics like CPU and memory usage is essential using tools like Prometheus. Logging activity across nodes is collected using Fluentd to help identify issues. Basic troubleshooting includes checking Pod logs and states, APIs to controllers, and ensuring sufficient resources.
Sanger, upcoming Openstack for Bio-informaticiansPeter Clapham
Delivery of a new Bio-informatics infrastructure at the Wellcome Trust Sanger Center. We include how to programatically create, manage and provide providence for images used both at Sanger and elsewhere using open source tools and continuous integration.
The document describes the DBE Execution Environment, which is a platform for small and medium-sized businesses across Europe. It consists of an integrated, distributed network of local digital ecosystems that allow businesses and local governments to cooperate by exchanging resources, applications, services, and knowledge dynamically. The environment is self-managed without a single owner or point of failure. It uses game theory, diffusion-reaction models, and other techniques to emerge organically. Services are deployed in containers called ServENTS that provide security, filtering, and other capabilities.
[Draft] Fast Prototyping with DPDK and eBPF in ContainernetAndrew Wang
This document discusses using Containernet as an environment for fast prototyping of network functions using DPDK and eBPF. Containernet allows creating virtual networks using Docker containers as hosts, providing scalability and access to debug tools. DPDK enables fast packet processing, while eBPF allows in-kernel packet processing without modifying the kernel. The document proposes demonstrating a prototype network in Containernet using DPDK to forward traffic to an application server, and eBPF filters on the server to process packets.
This document provides an overview and summary of OpenShift v3 and containers. It discusses how OpenShift v3 uses Docker containers and Kubernetes for orchestration instead of the previous "Gears" system. It also summarizes the key architectural changes in OpenShift v3, including using immutable Docker images, separating development and operations, and abstracting operational complexity.
Mathew Beane discusses strategies for optimizing and scaling Magento applications on clustered infrastructure. Some key points include:
- Using Puppetmaster to build out clusters with standard webnodes and database configurations.
- Magento supports huge stores and is very flexible and scalable. Redis is preferred over Memcache for caching.
- Important to have application optimization, testing protocols and deployment pipelines in place before scaling.
- Common components for scaling include load balancers, proxying web traffic, clustering Redis with Sentinel and Twemproxy, adding read servers and auto-scaling.
Nova compute and controller services provide virtual machines and coordinate nova services in OpenStack. Nova compute runs on each node and interacts with the hypervisor like KVM to launch VMs. The nova controller runs most nova services including the scheduler, which dispatches VM requests to nodes based on filters. Key components include the nova API, compute, and conductor services. Compute resources must be designed considering the number of processors, memory allocation, storage, resource pools, and over-commit ratios to optimize VM deployment in OpenStack.
The document describes a travel agency management system that offers the following key features:
- Integrated travel agents located directly in companies to make reservations and issue tickets.
- An electronic booking system that is IATA approved along with state-of-the-art technology.
- Dedicated and bilingual staff that provide personalized service and account management for corporate travel needs.
- One-stop shopping for all travel arrangements along with corporate agreements with airlines.
A presentation on how applying Cloud Architecture Patterns using Docker Swarm as orchestrator is possible to create reliable, resilient and scalable FIWARE platforms.
Wissbi is an open source toolset for building distributed event processing pipelines easily. It provides basic commands like wissbi-sub and wissbi-pub that allow receiving and sending messages. Filters can be written in any language and run in parallel as daemon processes configured through files. This allows constructing complex multi-stage data workflows. The ecosystem also includes tools like a log collector and metric collector that use Wissbi for transport. It aims to minimize operating effort through a simple design that relies mainly on filesystem operations and standard Unix tools and commands.
TechWiseTV Workshop: Open NX-OS and Devops with Puppet LabsRobb Boyd
Two incredible engineers: Shane Corban from Cisco and Carl Caum from Puppet Labs came together to be our guest experts for this workshop. See the demos in the replay at bit.ly/1lJQm3A
The document describes ViP2P, a system for distributed data management and querying in peer-to-peer networks using materialized views. It enables peers to pose queries over distributed data by rewriting queries using available views. The querying peer looks up view definitions, rewrites the query into a logical plan based on the views, generates an optimized physical plan, and executes it to return results. The system uses distributed hash tables, third-party libraries, and supports defining, indexing, materializing and querying distributed views across the peer-to-peer network.
The document discusses infrastructure as code (IAC) and its principles and categories. Some key points:
- IAC treats infrastructure like code by writing code to define, deploy, and update infrastructure. This allows infrastructure to be managed programmatically.
- Common categories of IAC include ad hoc scripts, configuration management tools like Ansible and Puppet, server templating tools like Packer, and server provisioning tools like Terraform.
- Benefits of IAC include automation, consistency, repeatability, versioning, validation, reuse, and allowing engineers to focus on code instead of manual tasks.
- AWS offers CloudFormation for provisioning AWS resources through templates. Other tools integrate with Cloud
Presentation from the first meetup of Kubernetes Pune - introduction to Kubernetes (https://www.meetup.com/Kubernetes-Pune/events/235689961)
Monitoring in Big Data Platform - Albert Lewandowski, GetInDataGetInData
Did you like it? Check out our blog to stay up to date: https://getindata.com/blog
The webinar was organized by GetinData on 2020. During the webinar we explaned the concept of monitoring and observability with focus on data analytics platforms.
Watch more here: https://www.youtube.com/watch?v=qSOlEN5XBQc
Whitepaper - Monitoring ang Observability for Data Platform: https://getindata.com/blog/white-paper-big-data-monitoring-observability-data-platform/
Speaker: Albert Lewandowski
Linkedin: https://www.linkedin.com/in/albert-lewandowski/
___
Getindata is a company founded in 2014 by ex-Spotify data engineers. From day one our focus has been on Big Data projects. We bring together a group of best and most experienced experts in Poland, working with cloud and open-source Big Data technologies to help companies build scalable data architectures and implement advanced analytics over large data sets.
Our experts have vast production experience in implementing Big Data projects for Polish as well as foreign companies including i.a. Spotify, Play, Truecaller, Kcell, Acast, Allegro, ING, Agora, Synerise, StepStone, iZettle and many others from the pharmaceutical, media, finance and FMCG industries.
https://getindata.com
Tips Tricks and Tactics with Cells and Scaling OpenStack - May, 2015Belmiro Moreira
The document discusses using cells in OpenStack to scale cloud infrastructure across multiple geographic locations. Key points include using cells to distribute OpenStack compute services around Australia, with over 6000 users, 700 hypervisors, and 30,000 cores spread across 8 sites and 14 cells. It also discusses strategies for operating, upgrading, and scheduling across multiple cells.
About 94% of AI Adopters are planning to use containers in the next 1 year. What’s driving this exponential growth? Faster time to deployment and Faster AI workload processing are the two major reasons. You can use GPUs in big data applications such as machine learning, data analytics, and genome sequencing. Docker containerization makes it easier for you to package and distribute applications. You can enable GPU support when using YARN on Docker containers. In this talk, I will demonstrate how Docker accelerates the AI workload development and deployment over the IoT Edge devices in efficient manner
Delivering Container-based Apps to IoT Edge devicesAjeet Singh Raina
I presented it during Dockercon. This talk was all about AI + Docker + IoT. Showcased how Docker app talk to Sensors, GPUs and Camera module and demo'ed how sensors data can be visualized over Grafana dashboard - all running on a IoT Edge device.
Presented Keynotes last weekend to Collabnix Community . Talk about trending topics, Kubernetes Statistics and community growth. It was fun talking about how exciting the community is and where are we heading next to.
Quantifying Your World with AI & Docker on the Edge | OSCONF 2020 JaipurAjeet Singh Raina
When Docker, IoT and AI Meet Together....
Highlights:
- Opendatacam is 100% Open Source solution
- Quantifies and tracks moving objects with Live Video Analysis
- Runs on Linux and CUDA GPU enabled hardware
- Runs completely on Docker Containers as well as support K3s.
Read the Full Story - https://collabnix.com/object-detection-with-yolo-using-docker-19-03-on-nvidia-jetson-nano/
Hear me LIVE on 10th October while I talk about Running Docker Container which uses GPU on Jetson Nano for Implementing Object Detection and Analytics.
IoET Conference is run by IoET Planet which is purely a collaborative platform. IoET Planet is an innovative platform to bring people together for the first time under a single umbrella in order to learn, collaborate and contribute.
It is an open community dedicated to IoT Research, innovation and applications. Today it is a blogging site & covers a wide range of topics, from embedded systems to data analytics and machine learning implementation, having the objective of enabling a vast array of services in areas like e-health, mobility, energy, manufacturing and agriculture, just to cite a few examples. These topics are coming from the community users and open for industry experts, IoT enthusiast and DevOps Engineers.
OSCONF is a platform built to connect & bring all leading Meetup communities, Cloud native experts, Evangelists & DevOps together under one roof. It is a non-profit community event which allows you to interact with community leaders, ambassadors & open community contributors.
Accelerate Your Automation Testing Effort using TestProject & Docker | Docker...Ajeet Singh Raina
Accelerate Your Test Automation using TestProject & Docker
A recording of a live webinar hosted on May 17th, 2020 - Learn from Docker Captain Ajeet Singh Raina how you can leverage TestProject Agents running in Docker containers, easily setup virtual labs & run tests in parallel.
- Introduction to TestProject
- Why TestProject for Automation?
- A Brief about TestProject Agents
- Why Docker?
- Why run TestProject Agents inside Docker containers?
- Live Demo
- Game with Prizes!
The Year 2020 is going to be super exciting for "Developers". As rightly stated by Scott Johnston, Chief Product Officer, Docker. Inc "Addressing the needs of developers has been core to Docker since its inception. Developers want the freedom to choose their own tools, the flexibility to quickly try new technologies and the ability to deploy their applications anywhere, independent of the underlying infrastructure,” Going forward, Docker Inc is going to focus on Docker Desktop, Docker Hub & Docker Compose's roles in the developer workflow for modern apps. Docker will be investing in expanding cloud services to enable developers to quickly discover technologies for use when building applications, to easily share these apps with teammates and the community, and to run apps frictionlessly on any Kubernetes endpoint, whether locally or in the cloud.
This year starting Jan 18th, Docker Bangalore Meetup community kickstarted 2020 with really exciting topics around Docker, Kubernetes & latest announcement from Kubecon.
Links to get Started with Docker/Kubernetes:
Collabnix - https://collabnix.com
DockerLabs - http://dockerlabs.collabnix.com
Kubelabs - http://kubelabs.collabnix.com
Kubetools - http://kubetools.collabnix.com
Gopherlabs - http://gopherlabs.collabnix.com
#Docker #Kubernetes #dockerlabs #cloudnative
Top 5 Helpful Tips to Grow Your Local Docker CommunityAjeet Singh Raina
The document provides 5 tips for growing a local Docker community: 1) be consistent with frequent meetup events, 2) choose a collaborative meetup platform, 3) actively engage on social media, 4) do not limit RSVP numbers but seek larger venues, and 5) create exciting event titles based on community feedback. It also provides two bonus tips: maintain meetup records on GitHub and collaborate on joint meetups with other groups initially.
Collabnix Community conduct webinar on regular basis. Swapnasagar Pradhan, an engineer from VISA delivered a talk on Traefik this January 11th 2020. Check this out.
Demystifying the Nuts & Bolts of Kubernetes ArchitectureAjeet Singh Raina
The document summarizes the architecture of Kubernetes. It uses an analogy of cargo ships and control ships to explain the different components. The master node components like the scheduler, ETCD cluster, and controller manager manage and monitor the worker nodes. The worker node components like Kubelet and kube-proxy run on each node and ensure pods and containers are running properly and can communicate. Pods are the basic building blocks that can contain one or more containers.
This document introduces Docker Compose, which allows defining and running multi-container Docker applications. It discusses that Docker Compose uses a YAML file to configure and run multi-service Docker apps. The 3 steps are to define services in a Dockerfile, define the app configuration in a Compose file, and run the containers with a single command. It also covers topics like networking, environment variables, and installing Docker Compose. Hands-on labs are provided to learn Compose through examples like WordPress.
I presented these slides during Docker Bangalore Meetup which happened on 23-Nov in DellEMC in Bangalore. These slides talks about the recent acquisition of Docker Enterprise by Mirantis and talks about what Docker is going to focus next on.
Prashansa Kulshrestha is currently pursuing B. Tech in Computer Science with Specialization in Cloud Computing and Virtualization Technology at UPES, Dehradun. She is Red Hat OpenShift Certified Engineer as well as RHCE too. She has experience working on Docker, Docker Compose & Docker Swarm. She is an active contributor of Docker Labs. She is expert in Python and well acquainted with AWS.
She recently implemented Log analytics tool using Elastic Stack and Docker and want to share her learning and experience with Collabnix Community.
Introduction to Docker Compose | Docker Intermediate WorkshopAjeet Singh Raina
Docker Compose allows users to define and run multi-container Docker applications. With Docker Compose, a YAML file is used to configure an application's services, and with a single command, all the services can be started from the configuration. Docker Compose is a three step process - services are defined in a Dockerfile, then in a Docker Compose file, and then run with docker-compose up. It supports volumes, networks, and environmental variables. Docker Compose can be used for development, testing, and production environments across different platforms.
Simplifying Real Time Data Analytics with Docker, IoT & CloudAjeet Singh Raina
Last Thursday, I travelled to Dehradun(around 1500 miles) for 1 day session on " The Pico project" as well as Docker workshop on Raspberry Pi. It was an amazing experience where I got chance to interact with the University students for the first time.
π0.5: a Vision-Language-Action Model with Open-World GeneralizationNABLAS株式会社
今回の資料「Transfusion / π0 / π0.5」は、画像・言語・アクションを統合するロボット基盤モデルについて紹介しています。
拡散×自己回帰を融合したTransformerをベースに、π0.5ではオープンワールドでの推論・計画も可能に。
This presentation introduces robot foundation models that integrate vision, language, and action.
Built on a Transformer combining diffusion and autoregression, π0.5 enables reasoning and planning in open-world settings.
ELectronics Boards & Product Testing_Shiju.pdfShiju Jacob
This presentation provides a high level insight about DFT analysis and test coverage calculation, finalizing test strategy, and types of tests at different levels of the product.
Concept of Problem Solving, Introduction to Algorithms, Characteristics of Algorithms, Introduction to Data Structure, Data Structure Classification (Linear and Non-linear, Static and Dynamic, Persistent and Ephemeral data structures), Time complexity and Space complexity, Asymptotic Notation - The Big-O, Omega and Theta notation, Algorithmic upper bounds, lower bounds, Best, Worst and Average case analysis of an Algorithm, Abstract Data Types (ADT)
its all about Artificial Intelligence(Ai) and Machine Learning and not on advanced level you can study before the exam or can check for some information on Ai for project
Passenger car unit (PCU) of a vehicle type depends on vehicular characteristics, stream characteristics, roadway characteristics, environmental factors, climate conditions and control conditions. Keeping in view various factors affecting PCU, a model was developed taking a volume to capacity ratio and percentage share of particular vehicle type as independent parameters. A microscopic traffic simulation model VISSIM has been used in present study for generating traffic flow data which some time very difficult to obtain from field survey. A comparison study was carried out with the purpose of verifying when the adaptive neuro-fuzzy inference system (ANFIS), artificial neural network (ANN) and multiple linear regression (MLR) models are appropriate for prediction of PCUs of different vehicle types. From the results observed that ANFIS model estimates were closer to the corresponding simulated PCU values compared to MLR and ANN models. It is concluded that the ANFIS model showed greater potential in predicting PCUs from v/c ratio and proportional share for all type of vehicles whereas MLR and ANN models did not perform well.
Analysis of reinforced concrete deep beam is based on simplified approximate method due to the complexity of the exact analysis. The complexity is due to a number of parameters affecting its response. To evaluate some of this parameters, finite element study of the structural behavior of the reinforced self-compacting concrete deep beam was carried out using Abaqus finite element modeling tool. The model was validated against experimental data from the literature. The parametric effects of varied concrete compressive strength, vertical web reinforcement ratio and horizontal web reinforcement ratio on the beam were tested on eight (8) different specimens under four points loads. The results of the validation work showed good agreement with the experimental studies. The parametric study revealed that the concrete compressive strength most significantly influenced the specimens’ response with the average of 41.1% and 49 % increment in the diagonal cracking and ultimate load respectively due to doubling of concrete compressive strength. Although the increase in horizontal web reinforcement ratio from 0.31 % to 0.63 % lead to average of 6.24 % increment on the diagonal cracking load, it does not influence the ultimate strength and the load-deflection response of the beams. Similar variation in vertical web reinforcement ratio leads to an average of 2.4 % and 15 % increment in cracking and ultimate load respectively with no appreciable effect on the load-deflection response.
Value Stream Mapping Worskshops for Intelligent Continuous SecurityMarc Hornbeek
This presentation provides detailed guidance and tools for conducting Current State and Future State Value Stream Mapping workshops for Intelligent Continuous Security.
Sorting Order and Stability in Sorting.
Concept of Internal and External Sorting.
Bubble Sort,
Insertion Sort,
Selection Sort,
Quick Sort and
Merge Sort,
Radix Sort, and
Shell Sort,
External Sorting, Time complexity analysis of Sorting Algorithms.
"Boiler Feed Pump (BFP): Working, Applications, Advantages, and Limitations E...Infopitaara
A Boiler Feed Pump (BFP) is a critical component in thermal power plants. It supplies high-pressure water (feedwater) to the boiler, ensuring continuous steam generation.
⚙️ How a Boiler Feed Pump Works
Water Collection:
Feedwater is collected from the deaerator or feedwater tank.
Pressurization:
The pump increases water pressure using multiple impellers/stages in centrifugal types.
Discharge to Boiler:
Pressurized water is then supplied to the boiler drum or economizer section, depending on design.
🌀 Types of Boiler Feed Pumps
Centrifugal Pumps (most common):
Multistage for higher pressure.
Used in large thermal power stations.
Positive Displacement Pumps (less common):
For smaller or specific applications.
Precise flow control but less efficient for large volumes.
🛠️ Key Operations and Controls
Recirculation Line: Protects the pump from overheating at low flow.
Throttle Valve: Regulates flow based on boiler demand.
Control System: Often automated via DCS/PLC for variable load conditions.
Sealing & Cooling Systems: Prevent leakage and maintain pump health.
⚠️ Common BFP Issues
Cavitation due to low NPSH (Net Positive Suction Head).
Seal or bearing failure.
Overheating from improper flow or recirculation.
"Boiler Feed Pump (BFP): Working, Applications, Advantages, and Limitations E...Infopitaara
Demystifying puppet
2. • Infrastructure-as-a-Code
• Puppet is an open source configuration management utility
• It is written in Ruby and released as free software under the GPL
• Built to be cross-platform
• A Declarative Language
• A Model Driven Architecture
10. • Talks about Resource Management when agents connect
• Handles the “How” by knowing how different platforms and OS manages certain type of
resources
• Each types has number of providers
• A Providers contains “how” of managing packages using a Package Management tool.
• When agents connect , Puppet uses a tool called “Facter”. To return the information about that
agent, including what OS is it running.
• Puppet chooses the appropriate package provider for that OS. Reports Success or Failure
11. • A System Inventory tool
• Returns facts about each agent(Hostname, IP address, OS and version information)
• These facts are gathered when agent runs.
• Facts are sent to the Puppet master, and automatically created as variable available to Puppet
• How to run facter?
• Facts are made available as variables that can be used in your puppet configuration
• Helps Puppet understand on how to manage particular resources in an agent.
13. • Resources - The core of the Puppet language is declaring resources. An Individual configuration
items
• Files – Physical files to serve to your agents
• Templates – Templates files that you can use to populate files
• Modules – Portable collection of resources. Reusable, sharable units of Puppet Code
• Classes – Modules can contain many Puppet classes. Groups of resource declarations and
conditional statements
• Manifests - Puppet code is saved in files called manifests, which are in turn stored in structured
directories called modules.
• Pre-built Puppet modules can be downloaded from the Puppet Forge, and most users will write at
least some of their own modules.
23. • All servers that are physical with 4 CPU, deploy ESX.
• All servers that are virtual with 1 CPU and 4GB of memory, deploy
CentOS, and handoff system to puppetmaster.cse.com for
management.
• All servers that are virtual with 32GB of memory, deploy Debian,
and handoff system to puppetmaster.cse.com for management.
24. - A Software tool for rapid provisioning of OS and Hypervisor
~ on both physical and virtual servers
- A policy-based bare-metal provisioning lets you inventory and manage the lifecycle of your
physical machines.
- Automatically discovers bare-metal hardware, dynamically configure operating systems
and/or hypervisors, and hand nodes off to PE for workload configuration
- Two major Components:
~ The Razor Server
(Ruby, MongoDB, Node.js)
~ The Razor Microkernel
(~20MB LInux Kernel, Facter, MCollective)
25. • - Discovery ( Tags, Matcher Rules)
• - Models(Defining OS Templates,..)
• - Policies(Rules that apply Models to Nodes based on Discovery)
• - Broker( Configuration Management)
26. When a new node appears, Razor discovers its characteristics by booting it with the Razor
microkernel and inventorying its facts.
The node is tagged based on its characteristics. Tags contain a match condition — a Boolean
expression that has access to the node’s facts and determines whether the tag should be applied to
the node or not.
29. • Install PE in Your Virtual Environment
• Install and Configure dnsmasq DHCP/TFTP Service
• Temporarily Disable SELinux to Enable PXE Boot
• Edit the dnsmasq Configuration File to Enable PXE Boot
• Install the Razor Server
– Load iPXE Software
– Verify the Razor Server
• Install and Set Up the Razor Client
• Setup Razor Provisioning
• Include Repos
• Include Brokers
• Include Tasks
• Create Policies
• Identify and Register Nodes
31. - A Tag consists of a unique name and rule
- The tag matches a node if evaluating it against the tag's facts results in true.
- The syntax for rule expressions is defined in lib/razor/matcher.rb
32. - Policies orchestrate repos, brokers, and tasks to tell Razor what bits to install, where to get the
bits, how they should be configured, and how to communicate between a node and PE.
- Policies contain a good deal of information, it’s handy to save them in a JSON file that you run when
you create the policy
Example: It should be applied to the first 20 nodes that have no more than two processors that boot
33. - Create a file called policy.json and copy the following template text into it:
- Edit the options in the policy.json template with information specific to your environment.
- Apply the policy by executing: razor create-policy --json policy.json.
34. Step -1 – A Fresh Razor with no new node.
Step-2 – Create a new VM. It retrieves DHCP IP and loads microkernel
45. Step-11- A New Node starts loading as per the policy specified.
46. Step-12 - Verify the node2 policy attached through puppet master
47. Step-13 - The New OS comes up , shows that it has been installed through Razor.
48. - Deploy version 1.2.3 on my application to all 3000 systems
- Deploy version 1.2.5rc2 of my application to all 340 development systems
- Restart the Apache service on all the systems in North America zones
- What systems are online right now?
- Run puppet on all systems, ensuring that at most 10 runs are happenings at once
- Upgrade the Hadoop version from 0.1 to 1.1 on all those 2500 nodes
49. - A framework to build server orchestration or parallel job execution systems
- Uses a Publish Subscribe Middleware philosophy ~ real time discovery of network resources
using meta data and not hostnames
~ a messaging pattern where senders of messages, called publishers, do not program the messages to be sent directly to specific
receivers, called subscribers. Instead, published messages are characterized into classes, without knowledge of what, if any, subscribers
there may be. Similarly, subscribers express interest in one or more classes, and only receive messages that are of interest, without
knowledge of what, if any, publishers there are. “
- Uses a broadcast paradigm for request distribution.
~ all servers get all requests at the same time, requests have filters attached and only servers matching the filter will act on requests.
There is no central asset database to go out of sync, the network is the only source of truth “
50. An MCollective client can send requests to any number of servers, using a security plugin to encode and sign the request
and a connector plugin to publish it.
It can also receive replies from servers, and format the response data for a user or some other system.
Example : mco command-line client
51. An mCollective server is a computer which can be controlled via mCollective.
Servers runs the mcollecitve daemon (mcollectived) and have any number of agent plugins installed
54. • An Abstract Cloud of Magic
• Clients and servers don’t communicate directly. They publish messages
to the middleware, and subscribe to messages they are interested in.
• Middleware system ~ knows how to route messages.
• External to MCollective, and everything that interacts directly with it is
pushed out into a connector plugin (which needs some limited
knowledge about the topology and configuration of the middleware).
55. Apache ActiveMQ
- an open-source message broker that runs on the JVM;
- Installed with a wrapper script and init script that allow it to be managed as a normal
OS service.
- MCollective talks to ActiveMQ using the Stomp protocol
- This is the main middleware recommended for use with Mcollective
- most well-tested option, its security features are powerful and flexible enough to suit
nearly all needs, and it can scale by clustering once a deployment gets too big (we
recommend ~800 nodes per ActiveMQ server as a maximum). Its main drawback is
that it can be frustrating to configure; to help mitigate that, we provide a detailed
ActiveMQ config reference in our own docs
•
56. RabbitMQ
• An open-source message broker written in Erlang; MCollective talks to
RabbitMQ using the Stomp protocol. Although it works well with
MCollective, it isn’t as thoroughly tested with it as ActiveMQ is, so if your
site has no preference, you should default to ActiveMQ.
• The RabbitMQ connector ships with MCollective’s core and is available by
default.
59. ~ Lets you define broadcast domains and configure a mcollective server to
belong to one or many of these domains.
60. $mco ping – communicates to all the puppet agents
61. - Make requests to your servers.
- Capable of interacting with any standard Remote Procedure Call (RPC) agent.
How it work?
i. Perform discovery against the network and discover 10 servers
ii. Send the request and then show a progress bar of the replies
iii. Show any results that were out of the ordinary
iv. Show some statistics
62. • You can request the status for a specific service
63. • Report of all your IBM hardware listing hostname, serial number, and product
name
64. Are you using the default Web Server?
Limitation of WEBrick, the default web server used to enable Puppet’s web services connectivity, is
essentially a reference implementation, and becomes unreliable as number of nodes increases.
Challenge-1 - Scaling the Transport
Possible Solution - increase the performance and potential number of possible master and agent
connections.
Challenge-2 – Scaling SSL
Possible Solution - we implement good management of the SSL certificates that secure the
connection between the master and the agent
• -
65. Approach
• Replacing WEBRick Ruby-based HTTP server with the Apache web server on a single Puppet master
system
• Extending the strategy to multiple Puppet master systems working behind a load balancer with Apache
Web Server/Nginx.
66. Running Puppet Master with Apache and Passenger
An Apache module that allows the embedding of Ruby applications, much like mod_php or mod_perl allow the
embedding of PHP and Perl applications.
For networks of one to two thousand Puppet managed nodes, a single Puppet master system running inside of
Apache with Passenger is often sufficient
Approach:
- Install Apache and Passenger
- Configure apache to handle SSL authentication and Verification of Puppet Agent
- Connect Apache to the Puppet Master
70. Front End HTTP Load Balancer
- Front-end Apache Virtual Host to authorize incoming Puppet Agent requests and handle SSL encryption
and decryption of the traffic
- It will terminate the SSL connection, be responsible for authenticating the puppet agent request and then
present this authentication information to the back-end Puppet master workers for authorization.
- All traffic between the front-end load balancer and the back-end Puppet master systems are unencrypted
and in plain text
71. Puppet Master Worker Configuration
• When running the Puppet master behind a load balancer, there will be multiple
Puppet master processes running on different hosts behind the load balancer.
• The load balancer will listen on the Puppet port of 8140. Incoming requests will be
dispatched to available back-end worker processes, (configures the Puppet CA and
workers all on the same host using unique TCP ports bound to the loopback
interface)
- .
75. - Synchronize the Puppet CA directory across all of the worker systems
- Make one worker system the active Puppet CA service and a second worker system the hot
standby Puppet CA service.