YugaByte DB on Kubernetes - An IntroductionYugabyte
This document summarizes YugaByte DB, a distributed SQL and NoSQL database. It discusses how YugaByte DB provides ACID transactions, strong consistency, and high performance at a planet scale. It also describes how to deploy YugaByte DB and an example e-commerce application called Yugastore on Kubernetes. The document outlines the database architecture and components, and provides steps to deploy the system and run a sample workload.
Distributed Database Architecture for GDPRYugabyte
The General Data Protection Regulation, often referred to as GDPR, came into effect on 25 May 2018 across the European Union. This regulation has implications on many global businesses, given the fines imposed if the organization is be found to be non-compliant. Making sure that the app architecture continues to ensure regulatory compliance is an on-going challenge for many businesses. This talk covesr some of the key requirements of GDPR that impact database architecture, what the inherent challenges are with requirements and how YugaByte DB can be used to implement these requirements.
Budapest Data Forum 2017 - BigQuery, Looker And Big Data Analytics At Petabyt...Rittman Analytics
As big data and data warehousing scale-up and move into the cloud, they’re increasingly likely to be delivered as services using distributed cloud query engines such as Google BigQuery, loaded using streaming data pipelines and queried using BI tools such as Looker. In this session the presenter will walk through how data modelling and query processing works when storing petabytes of customer event-level activity in a distributed data store and query engine like BigQuery, how data ingestion and processing works in an always-on streaming data pipeline, how additional services such as Google Natural Language API can be used to classify for sentiment and extract entity nouns from incoming unstructured data, and how BI tools such as Looker and Google Data Studio bring data discovery and business metadata layers to cloud big data analytics
SQL is a popular database language for modern applications, given its flexibility in modelling workloads and how widely it is understood by developers. However, most modern applications running in the clouds require fault tolerance, the ability to scale out and geographic data distribution of data. These are hard to achieve with traditional SQL databases, which is paving the way for distributed SQL databases.
Google Spanner is arguably the world's first truly distributed SQL database. Given its fully decentralized architecture, it delivers higher performance and availability for geo-distributed SQL workloads than other specialized transactional databases such as Amazon Aurora. Now, there are a number of open source derivatives of Google Spanner such as YugaByte DB, CockroachDB and TiDB. This talk will focus on the common architectural paradigms that these databases are built on (using YugaByte DB as an example). Learn about the concepts these databases leverage, how to evaluate if these will meet your needs and the questions to ask to differentiate among these databases.
This document provides an agenda and overview of a presentation on cloud data warehousing. The presentation discusses data challenges companies face today with large and diverse data sources, and how a cloud data warehouse can help address these challenges by providing unlimited scalability, flexibility, and lower costs. It introduces Snowflake as a first cloud data warehouse built for the cloud, with features like separation of storage and compute, automatic query optimization, and built-in security and encryption. Other cloud data warehouse offerings like Amazon Redshift are also briefly discussed.
Webinar: High Performance MongoDB Applications with IBM POWER8MongoDB
Innovative companies are building Internet of Things, mobile, content management, single view, and big data apps on top of MongoDB. In this session, we'll explore how the IBM POWER8 platform brings new levels of performance and ease of configuration to these solutions which already benefit from easier and faster design and development using MongoDB.
This document outlines an agenda for a 90-minute workshop on Snowflake. The agenda includes introductions, an overview of Snowflake and data warehousing, demonstrations of how users utilize Snowflake, hands-on exercises loading sample data and running queries, and discussions of Snowflake architecture and capabilities. Real-world customer examples are also presented, such as a pharmacy building new applications on Snowflake and an education company using it to unify their data sources and achieve a 16x performance improvement.
- MongoDB is a document database management system that is recognized as a leader by Gartner. It has over 520 employees, 2500+ customers, and offices globally.
- MongoDB ranked 4th in database mindshare according to DB-Engines. It has seen 172% growth in the last 20 months.
- Several companies such as a quantitative investment manager, an insurance company, a telecommunications company, and an ecommerce company migrated their systems to MongoDB and saw benefits like 100x faster data retrieval, 50% lower costs, and being able to build applications faster.
Webinar: Faster Big Data Analytics with MongoDBMongoDB
Learn how to leverage MongoDB and Big Data technologies to derive rich business insight and build high performance business intelligence platforms. This presentation includes:
- Uncovering Opportunities with Big Data analytics
- Challenges of real-time data processing
- Best practices for performance optimization
- Real world case study
This presentation was given in partnership with CIGNEX Datamatics.
Webinar: Enterprise Trends for Database-as-a-ServiceMongoDB
Two complementary trends are particularly strong in enterprise IT today: MongoDB itself, and the movement of infrastructure, platform, and software to as-a-service models. Being designed from the start to work in cloud deployments, MongoDB is a natural fit.
Learn how your enterprise can create its own MongoDB service offering, combining the advantages of MongoDB and cloud for agile, nearly-instantaneous deployments. Ease your operations workload by centralizing your points for enforcement, standardize best policies, and enable elastic scalability.
We will provide you with an enterprise planning outline which incorporates needs and value for stakeholders across operations, development, and business. We will cover accounting, chargeback integration, and quantification of benefits to the enterprise (such as standardizing best practices, creating elastic architecture, and reducing database maintenance costs).
Advanced Reporting and ETL for MongoDB: Easily Build a 360-Degree View of You...MongoDB
The document discusses Pentaho's analytics and ETL solutions for MongoDB. It provides an overview of Pentaho Company and its platform for unified business analytics and data integration. It then outlines how Pentaho can be used to build a 360-degree view of customers by extracting, transforming and loading data from source systems into MongoDB and performing analytics and reporting on the MongoDB data. It demonstrates these capabilities with examples and screenshots.
Join CIGNEX Datamatics, Alfresco’s Global Platinum Partner, as they share the case study experience of a leading global online university. Together we’ll take a close look at their document management and web portal solution and their integrations with Alfresco ECM, Liferay Portal and Moodle Learning Management System.
Actionable Insights with AI - Snowflake for Data ScienceHarald Erb
Talk @ ScaleUp 360° AI Infrastructures DACH, 2021: Data scientists spend 80% and more of their time searching for and preparing data. This talk explains Snowflake’s Platform capabilities like near-unlimited data storage and instant and near-infinite compute resources and how the platform can be used to seamlessly integrate and support the machine learning libraries and tools data scientists rely on.
As Twitch grew, both the amount of data we received and the number of employees interested in the data grew rapidly. In order to continue empowering decision making as we scaled, we turned to using Druid and Imply to provide self service analytics to both our technical and non technical staff allowing them to drill into high level metrics in lieu of reading generated reports.
In this talk, learn how Twitch implemented a common analytics platform for the needs of many different teams supporting hundreds of users, thousands of queries, and ~5 billion events each day. This session will explain our Druid architecture in detail, including:
-The end-to-end architecture deployed on Amazon that includes Kinesis, RDS, S3, Druid, Pivot and Tableau
-How the data is brought together to deliver a unified view of live customer engagement and historical trends
-Operational best practices we learnt scaling Druid
-An example walk through using the platform
Enterprise Reporting with MongoDB and JasperSoftMongoDB
The document provides an agenda and overview for a briefing on using MongoDB and JasperSoft for enterprise reporting. It discusses MongoDB's data model of flexible documents, query model with rich queries and analytics functions. It then outlines use cases of JasperSoft and MongoDB together for data hubbing from various sources, real-time analytics dashboards, and examples of customers using the integrated solution.
OBJETIVOS DO EVENTO
Fortalecer os estudos na área de Business Intelligence;
Promover o desenvolvimento de técnicas, metodologias e interfaces junto a comunidade;
Gerar interação entre estudantes, profissionais e empresas aumentado a qualidade do Networking.
Here I talk about examples and use cases for Big Data & Big Data Analytics and how we accomplished massive-scale sentiment, campaign and marketing analytics for Razorfish using a collecting of database, Big Data and analytics technologies.
Google BigQuery is Google's fully managed big data analytics service that allows users to analyze very large datasets. It offers a fast and easy to use service with no infrastructure to manage. Developers can stream up to 100,000 rows of data per second for near real-time analysis. BigQuery bills users per project on a pay-as-you-go model, with the first 1TB of data processed each month free of charge.
AWS Summit Singapore 2019 | Snowflake: Your Data. No LimitsAWS Summits
This document discusses Snowflake, a cloud data platform. It describes Snowflake's mission to enable organizations to be data-driven. It outlines problems with traditional data architectures like complexity, limited scalability, inability to consolidate data, and rigid costs. Snowflake's solution is a cloud-native data warehouse delivered as a service that offers instant elasticity, end-to-end security, and the ability to query structured and semi-structured data using SQL. Key benefits of Snowflake include supporting any scale of data, users and workloads; paying only for resources used; and providing simplicity, scalability, flexibility and elasticity.
Smartsheet’s Transition to Snowflake and Databricks: The Why and Immediate Im...Databricks
Join this session to hear why Smartsheet decided to transition from their entirely SQL-based system to Snowflake and Databricks, and learn how that transition has made an immediate impact on their team, company and customer experience through enabling faster, informed data decisions.
The document discusses MongoDB operations for developers, including its data model, use of replication for high availability, sharding for scalability, and deployment architectures. It also covers MongoDB's philosophy, benefits of its document model, how replica sets provide self-healing and failure recovery, and security features available in MongoDB Enterprise.
in this presentation we go through the differences and similarities between Redshift and BigQuery. It was presented during the Athens Big Data meetup May 2017.
MongoDB .local London 2019: MongoDB Atlas Data Lake Technical Deep DiveMongoDB
This document summarizes MongoDB Atlas Data Lake, a new service that allows customers to access, query, and analyze long-term data stored in AWS S3 buckets. It implements MongoDB's query language and security model to provide a familiar interface for working with structured data in object storage. The service is read-only, distributed, and optimized to handle queries over vast amounts of data efficiently using MongoDB's aggregation engine. Customers maintain full control over their data and how it is configured and accessed.
Architecting Snowflake for High Concurrency and High PerformanceSamanthaBerlant
Cloud Data Warehousing juggernaut Snowflake has raced out ahead of the pack to deliver a data management platform from which a wealth of new analytics can be run. Using Snowflake as a traditional data warehouse has some obvious cost advantages over a hardware solution. But the real value of Snowflake as a data platform lies in its ability to support a high-concurrency analytics platform using Kyligence Cloud, powered by Apache Kylin.
In this presentation, Senior Solutions Architect Robert Hardaway will describe a modern data service architecture using precomputation and distributed indexes to provide interactive analytics to hundreds or even thousands of users running against very large Snowflake datasets (TBs to PBs).
During this presentation, Infusion and MongoDB shared their mainframe optimization experiences and best practices. These have been gained from working with a variety of organizations, including a case study from one of the world’s largest banks. MongoDB and Infusion bring a tested approach that provides a new way of modernizing mainframe applications, while keeping pace with the demand for new digital services.
Vivint Smart Home's journey with Snowflake and migrating from SQL Server. We describe how we have setup snowflake from a people, process, and technology perspective.
YugaByte DB is a transactional database that provides SQL and NoSQL interfaces in a single platform. It was created to address the complexity of building applications using separate SQL and NoSQL databases. YugaByte DB integrates with PKS to enable deployment on Kubernetes clusters. The presentation provides an overview of YugaByte DB's architecture and capabilities, demonstrates its integration with PKS, and discusses several real-world use cases.
YugaByte DB Internals - Storage Engine and Transactions Yugabyte
This document introduces YugaByte DB, a high-performance, distributed, transactional database. It is built to scale horizontally on commodity servers across data centers for mission-critical applications. YugaByte DB uses a transactional document store based on RocksDB, Raft-based replication for resilience, and automatic sharding and rebalancing. It supports ACID transactions across documents, provides APIs compatible with Cassandra and Redis, and is open source. The architecture is designed for high performance, strong consistency, and cloud-native deployment.
How YugaByte DB Implements Distributed PostgreSQLYugabyte
Building applications on PostgreSQL that require automatic data sharding and replication, fault tolerance, distributed transactions and geographic data distribution has been hard. In this 3 hour workshop, we will look at how to do this using a real-world example running on top of YugaByte DB, a distributed database that is fully wire-compatible with PostgreSQL and NoSQL APIs (Apache Cassandra and Redis). We will look at the design and architecture of YugaByte DB and how it reuses the PostgreSQL codebase to achieve full API compatibility. YugaByte DB support for PostgreSQL includes most data types, queries, stored procedures, etc. We will also take a look at how to build applications that are planet scale (requiring geographic distribution of data) and how to run them in cloud-native environments (for example, Kubernetes, hybrid or multi-cloud deployments).
Webinar: Faster Big Data Analytics with MongoDBMongoDB
Learn how to leverage MongoDB and Big Data technologies to derive rich business insight and build high performance business intelligence platforms. This presentation includes:
- Uncovering Opportunities with Big Data analytics
- Challenges of real-time data processing
- Best practices for performance optimization
- Real world case study
This presentation was given in partnership with CIGNEX Datamatics.
Webinar: Enterprise Trends for Database-as-a-ServiceMongoDB
Two complementary trends are particularly strong in enterprise IT today: MongoDB itself, and the movement of infrastructure, platform, and software to as-a-service models. Being designed from the start to work in cloud deployments, MongoDB is a natural fit.
Learn how your enterprise can create its own MongoDB service offering, combining the advantages of MongoDB and cloud for agile, nearly-instantaneous deployments. Ease your operations workload by centralizing your points for enforcement, standardize best policies, and enable elastic scalability.
We will provide you with an enterprise planning outline which incorporates needs and value for stakeholders across operations, development, and business. We will cover accounting, chargeback integration, and quantification of benefits to the enterprise (such as standardizing best practices, creating elastic architecture, and reducing database maintenance costs).
Advanced Reporting and ETL for MongoDB: Easily Build a 360-Degree View of You...MongoDB
The document discusses Pentaho's analytics and ETL solutions for MongoDB. It provides an overview of Pentaho Company and its platform for unified business analytics and data integration. It then outlines how Pentaho can be used to build a 360-degree view of customers by extracting, transforming and loading data from source systems into MongoDB and performing analytics and reporting on the MongoDB data. It demonstrates these capabilities with examples and screenshots.
Join CIGNEX Datamatics, Alfresco’s Global Platinum Partner, as they share the case study experience of a leading global online university. Together we’ll take a close look at their document management and web portal solution and their integrations with Alfresco ECM, Liferay Portal and Moodle Learning Management System.
Actionable Insights with AI - Snowflake for Data ScienceHarald Erb
Talk @ ScaleUp 360° AI Infrastructures DACH, 2021: Data scientists spend 80% and more of their time searching for and preparing data. This talk explains Snowflake’s Platform capabilities like near-unlimited data storage and instant and near-infinite compute resources and how the platform can be used to seamlessly integrate and support the machine learning libraries and tools data scientists rely on.
As Twitch grew, both the amount of data we received and the number of employees interested in the data grew rapidly. In order to continue empowering decision making as we scaled, we turned to using Druid and Imply to provide self service analytics to both our technical and non technical staff allowing them to drill into high level metrics in lieu of reading generated reports.
In this talk, learn how Twitch implemented a common analytics platform for the needs of many different teams supporting hundreds of users, thousands of queries, and ~5 billion events each day. This session will explain our Druid architecture in detail, including:
-The end-to-end architecture deployed on Amazon that includes Kinesis, RDS, S3, Druid, Pivot and Tableau
-How the data is brought together to deliver a unified view of live customer engagement and historical trends
-Operational best practices we learnt scaling Druid
-An example walk through using the platform
Enterprise Reporting with MongoDB and JasperSoftMongoDB
The document provides an agenda and overview for a briefing on using MongoDB and JasperSoft for enterprise reporting. It discusses MongoDB's data model of flexible documents, query model with rich queries and analytics functions. It then outlines use cases of JasperSoft and MongoDB together for data hubbing from various sources, real-time analytics dashboards, and examples of customers using the integrated solution.
OBJETIVOS DO EVENTO
Fortalecer os estudos na área de Business Intelligence;
Promover o desenvolvimento de técnicas, metodologias e interfaces junto a comunidade;
Gerar interação entre estudantes, profissionais e empresas aumentado a qualidade do Networking.
Here I talk about examples and use cases for Big Data & Big Data Analytics and how we accomplished massive-scale sentiment, campaign and marketing analytics for Razorfish using a collecting of database, Big Data and analytics technologies.
Google BigQuery is Google's fully managed big data analytics service that allows users to analyze very large datasets. It offers a fast and easy to use service with no infrastructure to manage. Developers can stream up to 100,000 rows of data per second for near real-time analysis. BigQuery bills users per project on a pay-as-you-go model, with the first 1TB of data processed each month free of charge.
AWS Summit Singapore 2019 | Snowflake: Your Data. No LimitsAWS Summits
This document discusses Snowflake, a cloud data platform. It describes Snowflake's mission to enable organizations to be data-driven. It outlines problems with traditional data architectures like complexity, limited scalability, inability to consolidate data, and rigid costs. Snowflake's solution is a cloud-native data warehouse delivered as a service that offers instant elasticity, end-to-end security, and the ability to query structured and semi-structured data using SQL. Key benefits of Snowflake include supporting any scale of data, users and workloads; paying only for resources used; and providing simplicity, scalability, flexibility and elasticity.
Smartsheet’s Transition to Snowflake and Databricks: The Why and Immediate Im...Databricks
Join this session to hear why Smartsheet decided to transition from their entirely SQL-based system to Snowflake and Databricks, and learn how that transition has made an immediate impact on their team, company and customer experience through enabling faster, informed data decisions.
The document discusses MongoDB operations for developers, including its data model, use of replication for high availability, sharding for scalability, and deployment architectures. It also covers MongoDB's philosophy, benefits of its document model, how replica sets provide self-healing and failure recovery, and security features available in MongoDB Enterprise.
in this presentation we go through the differences and similarities between Redshift and BigQuery. It was presented during the Athens Big Data meetup May 2017.
MongoDB .local London 2019: MongoDB Atlas Data Lake Technical Deep DiveMongoDB
This document summarizes MongoDB Atlas Data Lake, a new service that allows customers to access, query, and analyze long-term data stored in AWS S3 buckets. It implements MongoDB's query language and security model to provide a familiar interface for working with structured data in object storage. The service is read-only, distributed, and optimized to handle queries over vast amounts of data efficiently using MongoDB's aggregation engine. Customers maintain full control over their data and how it is configured and accessed.
Architecting Snowflake for High Concurrency and High PerformanceSamanthaBerlant
Cloud Data Warehousing juggernaut Snowflake has raced out ahead of the pack to deliver a data management platform from which a wealth of new analytics can be run. Using Snowflake as a traditional data warehouse has some obvious cost advantages over a hardware solution. But the real value of Snowflake as a data platform lies in its ability to support a high-concurrency analytics platform using Kyligence Cloud, powered by Apache Kylin.
In this presentation, Senior Solutions Architect Robert Hardaway will describe a modern data service architecture using precomputation and distributed indexes to provide interactive analytics to hundreds or even thousands of users running against very large Snowflake datasets (TBs to PBs).
During this presentation, Infusion and MongoDB shared their mainframe optimization experiences and best practices. These have been gained from working with a variety of organizations, including a case study from one of the world’s largest banks. MongoDB and Infusion bring a tested approach that provides a new way of modernizing mainframe applications, while keeping pace with the demand for new digital services.
Vivint Smart Home's journey with Snowflake and migrating from SQL Server. We describe how we have setup snowflake from a people, process, and technology perspective.
YugaByte DB is a transactional database that provides SQL and NoSQL interfaces in a single platform. It was created to address the complexity of building applications using separate SQL and NoSQL databases. YugaByte DB integrates with PKS to enable deployment on Kubernetes clusters. The presentation provides an overview of YugaByte DB's architecture and capabilities, demonstrates its integration with PKS, and discusses several real-world use cases.
YugaByte DB Internals - Storage Engine and Transactions Yugabyte
This document introduces YugaByte DB, a high-performance, distributed, transactional database. It is built to scale horizontally on commodity servers across data centers for mission-critical applications. YugaByte DB uses a transactional document store based on RocksDB, Raft-based replication for resilience, and automatic sharding and rebalancing. It supports ACID transactions across documents, provides APIs compatible with Cassandra and Redis, and is open source. The architecture is designed for high performance, strong consistency, and cloud-native deployment.
How YugaByte DB Implements Distributed PostgreSQLYugabyte
Building applications on PostgreSQL that require automatic data sharding and replication, fault tolerance, distributed transactions and geographic data distribution has been hard. In this 3 hour workshop, we will look at how to do this using a real-world example running on top of YugaByte DB, a distributed database that is fully wire-compatible with PostgreSQL and NoSQL APIs (Apache Cassandra and Redis). We will look at the design and architecture of YugaByte DB and how it reuses the PostgreSQL codebase to achieve full API compatibility. YugaByte DB support for PostgreSQL includes most data types, queries, stored procedures, etc. We will also take a look at how to build applications that are planet scale (requiring geographic distribution of data) and how to run them in cloud-native environments (for example, Kubernetes, hybrid or multi-cloud deployments).
YugaByte DB - "Designing a Distributed Database Architecture for GDPR Complia...Jimmy Guerrero
Join Karthik Ranganathan (YugaByte CTO) for an in-depth technical webinar to understand how developers and administrators alike, can design systems that enable users to control the sharing and protection of their personal data so that it complies with GDPR. Topics covered include schema design, data partitioning, encryption and replication. Karthik will draw on his experience helping scale Facebook's Messenger and Inbox Search along with real-world implementations which make use of YugaByte DB.
Docker containers are great for running stateless microservices, but what about stateful applications such as databases and persistent queues? Kubernetes provides the StatefulSets controller for such applications that have to manage data in some form of persistent storage. While StatefulSets is a great start, a lot more goes into ensuring high performance, data durability and high availability for stateful apps in Kubernetes. Following are 5 best practices that developers and operations engineers should be aware of.
1. Ensure high performance with local persistent volumes and pod anti-affinity rules.
2. Achieve data resilience with auto-failover and multi-zone pod scheduling.
3. Integrate StatefulSet services with other application services through NodePorts & LoadBalancer services.
4. Run Day 2 operations such as monitoring, elastic scaling, capacity re-sizing, backups with caution.
5. Automate operations through Kubernetes Operators that extend the StatefulSets controller.
We will demonstrate how to run a complete E-Commerce application powered by YugaByte DB, when all services are deployed in Kubernetes.
Scale Transactional Apps Across Multiple Regions with Low LatencyYugabyte
User-facing transactional apps in verticals such as Retail, Finance and SaaS are increasingly moving from a single-region, monolithic architecture to a multi-region, cloud-native architecture. Enhancing customer satisfaction with low latency access, protecting data through geo-redundancy and satisfying compliance requirements such as GDPR are some of the major drivers for this move. Unfortunately, the DB tier powering the above apps has remained as a high-latency, hard-to-scale master-slave RDBMS for a long time. Multi-master deployments as well as the use of a separate NoSQL DB for multi-region data distribution are simply band-aids to this problem and do not deliver the desired business outcomes.
This talk shows how to use YugaByte DB to scale transactional apps across multiple regions with low latency.
Make your data fly - Building data platform in AWSKimmo Kantojärvi
This document summarizes a presentation on building a data platform in AWS. It discusses the architectural evolution from on-premise data warehouses to cloud-based data lakes and platforms. It provides examples of using AWS services like EMR, Redshift, Airflow and visualization tools. It also covers best practices for data modeling, performance optimization, security and DevOps approaches.
Modernizing Global Shared Data Analytics Platform and our Alluxio JourneyAlluxio, Inc.
Data Orchestration Summit 2020 organized by Alluxio
https://www.alluxio.io/data-orchestration-summit-2020/
Modernizing Global Shared Data Analytics Platform and our Alluxio Journey
Sandipan Chakraborty, Director of Engineering (Rakuten)
About Alluxio: alluxio.io
Engage with the open source community on slack: alluxio.io/slack
Webinar slides: Free Monitoring (on Steroids) for MySQL, MariaDB, PostgreSQL ...Severalnines
Traditional server monitoring tools are not built for modern distributed database architectures. Let’s face it, most production databases today run in some kind of high availability setup - from simpler master-slave replication to multi-master clusters fronted by redundant load balancers. Operations teams deal with dozens, often hundreds of services that make up the database environment.
This is why we built ClusterControl - to address modern, highly distributed database setups based on replication or clustering. We wanted something that could provide a systems view of all the components of a distributed cluster, including load balancers.
Watch this replay of a webinar on free database monitoring using ClusterControl Community Edition. We show you how to monitor all your MySQL, MariaDB, PostgreSQL and MongoDB systems from a single point of control - whether they are deployed as Galera Clusters, sharded clusters or replication setups across on-prem and cloud data centers. We also see how to use Advisors in order to improve performance.
AGENDA
- Requirements for monitoring distributed database systems
- Cloud-based vs On-prem monitoring solutions
- Agent-based vs Agentless monitoring
- Deepdive into ClusterControl Community Edition
- Architecture
- Metrics Collection
- Trending
- Dashboards
- Queries
- Performance Advisors
- Other features available to Community users
SPEAKER
Bartlomiej Oles is a MySQL and Oracle DBA, with over 15 years experience in managing highly available production systems at IBM, Nordea Bank, Acxiom, Lufthansa, and other Fortune 500 companies. In the past five years, his focus has been on building and applying automation tools to manage multi-datacenter database environments.
Comparing three data ingestion approaches where Apache Kafka integrates with ...HostedbyConfluent
Using Kafka to stream data into TigerGraph, a distributed graph database, is a common pattern in our customers’ data architecture. We have seen the integration in three different layers around TigerGraph’s data flow architecture, and many key use case areas such as customer 360, entity resolution, fraud detection, machine learning, and recommendation engine. Firstly, TigerGraph’s internal data ingestion architecture relies on Kafka as an internal component. Secondly, TigerGraph has a builtin Kafka Loader, which can connect directly with an external Kafka cluster for data streaming. Thirdly, users can use an external Kafka cluster to connect other cloud data sources to TigerGraph cloud database solutions through the built-in Kafka Loader feature. In this session, we will present the high-level architecture in three different approaches and demo the data streaming process.
New enhancements for security and usability in EDB 13EDB
This document provides an overview of new enhancements in EDB 13 for security, usability and compatibility. Key highlights include improvements to Postgres Enterprise Manager for managing very large databases, enhanced security features like channel binding for authentication, and improved compatibility for Oracle migrations through features like automatic partitioning and Oracle compatible functions. It also outlines new capabilities in PostgreSQL 13 like parallel vacuuming and security tools, as well as enhancements to EDB tools for high availability, backup/recovery and Oracle compatibility.
During this webinar, we will review best practices and lessons learned from working with large and mid-size companies on their deployment of PostgreSQL. We will explore the practices that helped industry leaders move through these stages quickly, and get as much value out of PostgreSQL as possible without incurring undue risk.
We have identified a set of levers that companies can use to accelerate their success with PostgreSQL:
- Application Tiering
- Collaboration between DBAs and Development Teams
- Evangelizing
- Standardization and Automation
- Balance of Migration and New Development
Managing Postgres at Scale With Postgres Enterprise ManagerEDB
Dave Page shows you how to manage large scale Postgres deployments and highlights how Postgres Enterprise Manager can be used for monitoring, alerting and administration of your Postgres estate - no matter where it is deployed.
Big Data Pipeline for Analytics at Scale @ FIT CVUT 2014Jaroslav Gergic
The recent boom in big data processing and democratization of the big data space has been enabled by the fact that most of the concepts originated in the research labs of companies such as Google, Amazon, Yahoo and Facebook are now available as open source. Technologies such as Hadoop, Cassandra let businesses around the world to become more data driven and tap into their massive data feeds to mine valuable insights.
At the same time, we are still at a certain stage of the maturity curve of these new big data technologies and of the entire big data technology stack. Many of the technologies originated from a particular use case and attempts to apply them in a more generic fashion are hitting the limits of their technological foundations. In some areas, there are several competing technologies for the same set of use cases, which increases risks and costs of big data implementations.
We will show how GoodData solves the entire big data pipeline today, starting from raw data feeds all the way up to actionable business insights. All this provided as a hosted multi-tenant environment letting its customers to solve their particular analytical use case or many analytical use cases for thousands of their customers all using the same platform and tools while abstracting them away from the technological details of the big data stack.
An overview of reference architectures for PostgresEDB
EDB Reference Architectures are designed to help new and existing users alike to quickly design a deployment architecture that suits their needs. They can be used as either the blueprint for a deployment, or as the basis for a design that enhances and extends the functionality and features offered.
Add-on architectures allow users to easily extend their core database server deployment to add additional features and functionality "building block" style.
In this webinar, we will review the following architectures:
- Single Node
- Multi Node with Asynchronous Replication
- Multi Node with Synchronous Replication
- Add-on Architectures
An overview of reference architectures for PostgresEDB
EDB Reference Architectures are designed to help new and existing users alike to quickly design a deployment architecture that suits their needs. They can be used as either the blueprint for a deployment, or as the basis for a design that enhances and extends the functionality and features offered.
Add-on architectures allow users to easily extend their core database server deployment to add additional features and functionality "building block" style.
In this webinar, we will review the following architectures:
- Single Node
- Multi Node with Asynchronous Replication
- Multi Node with Synchronous Replication
- Add-on Architectures
Speaker:
Michael Willer
Sales Engineer, EDB
Gimel at Dataworks Summit San Jose 2018Romit Mehta
Gimel is PayPal's data platform that provides a unified interface for accessing and analyzing data across different data stores and processing engines. The presentation provides an overview of Gimel, including PayPal's analytics ecosystem, the challenges Gimel addresses around data access and application lifecycle, and a demo of how Gimel simplifies a flights cancelled use case. It also discusses Gimel's open source journey and integration with ecosystems like Spark and Jupyter notebooks.
Gimel Data Platform is an analytics platform developed by PayPal that aims to simplify data access and analysis. The presentation provides an overview of Gimel, including PayPal's analytics ecosystem, the challenges Gimel addresses in data access and application lifecycle management, a demo of a sample flights cancelled use case using Gimel, and PayPal's plans to open source Gimel.
Learn what's new in EDB Postgres 11. This update includes a refreshed version of EDB Postgres Advanced Server, which is built on and enhances the capabilities of open source PostgreSQL 11 with new data redaction capabilities, autonomous transaction commands, and performance diagnostics.
Webinar agenda:
- An intro to EDB Postgres, including BART, EFM, and containers
- What's new with EDB Postgres 11
- Brief overview and demo of PEM 7
Don’t miss this opportunity to hear from some of the top Postgres contributors!
1. The document discusses Project Geode, an open source distributed in-memory database for big data applications. It provides scale-out performance, consistent operations across nodes, high availability, powerful developer features, and easy administration of distributed nodes.
2. The document outlines Geode's architecture and roadmap. It also discusses why the project is being open sourced under Apache and describes some key use cases and customers of Geode.
3. The presentation includes a demo of Geode's capabilities including partitioning, queries, indexing, colocation, and transactions.
What AI Means For Your Product Strategy And What To Do About ItVMware Tanzu
The document summarizes Matthew Quinn's presentation on "What AI Means For Your Product Strategy And What To Do About It" at Denver Startup Week 2023. The presentation discusses how generative AI could impact product strategies by potentially solving problems companies have ignored or allowing competitors to create new solutions. Quinn advises product teams to evaluate their strategies and roadmaps, ensure they understand user needs, and consider how AI may change the problems being addressed. He provides examples of how AI could influence product development for apps in home organization and solar sales. Quinn concludes by urging attendees not to ignore AI's potential impacts and to have hard conversations about emerging threats and opportunities.
Make the Right Thing the Obvious Thing at Cardinal Health 2023VMware Tanzu
This document discusses the evolution of internal developer platforms and defines what they are. It provides a timeline of how technologies like infrastructure as a service, public clouds, containers and Kubernetes have shaped developer platforms. The key aspects of an internal developer platform are described as providing application-centric abstractions, service level agreements, automated processes from code to production, consolidated monitoring and feedback. The document advocates that internal platforms should make the right choices obvious and easy for developers. It also introduces Backstage as an open source solution for building internal developer portals.
Enhancing DevEx and Simplifying Operations at ScaleVMware Tanzu
Cardinal Health introduced Tanzu Application Service in 2016 and set up foundations for cloud native applications in AWS and later migrated to GCP in 2018. TAS has provided Cardinal Health with benefits like faster development of applications, zero downtime for critical applications, hosting over 5,000 application instances, quicker patching for security vulnerabilities, and savings through reduced lead times and staffing needs.
Dan Vega discussed upcoming changes and improvements in Spring including Spring Boot 3, which will have support for JDK 17, Jakarta EE 9/10, ahead-of-time compilation, improved observability with Micrometer, and Project Loom's virtual threads. Spring Boot 3.1 additions were also highlighted such as Docker Compose integration and Spring Authorization Server 1.0. Spring Boot 3.2 will focus on embracing virtual threads from Project Loom to improve scalability of web applications.
Platforms, Platform Engineering, & Platform as a ProductVMware Tanzu
This document discusses building platforms as products and reducing developer toil. It notes that platform engineering now encompasses PaaS and developer tools. A quote from Mercedes-Benz emphasizes building platforms for developers, not for the company itself. The document contrasts reactive, ticket-driven approaches with automated, self-service platforms and products. It discusses moving from considering platforms as a cost center to experts that drive business results. Finally, it provides questions to identify sources of developer toil, such as issues with workstation setup, running software locally, integration testing, committing changes, and release processes.
This document provides an overview of building cloud-ready applications in .NET. It defines what makes an application cloud-ready, discusses common issues with legacy applications, and recommends design patterns and practices to address these issues, including loose coupling, high cohesion, messaging, service discovery, API gateways, and resiliency policies. It includes code examples and links to additional resources.
Dan Vega discussed new features and capabilities in Spring Boot 3 and beyond, including support for JDK 17, Jakarta EE 9, ahead-of-time compilation, observability with Micrometer, Docker Compose integration, and initial support for Project Loom's virtual threads in Spring Boot 3.2 to improve scalability. He provided an overview of each new feature and explained how they can help Spring applications.
Spring Cloud Gateway - SpringOne Tour 2023 Charles Schwab.pdfVMware Tanzu
Spring Cloud Gateway is a gateway that provides routing, security, monitoring, and resiliency capabilities for microservices. It acts as an API gateway and sits in front of microservices, routing requests to the appropriate microservice. The gateway uses predicates and filters to route requests and modify requests and responses. It is lightweight and built on reactive principles to enable it to scale to thousands of routes.
This document appears to be from a VMware Tanzu Developer Connect presentation. It discusses Tanzu Application Platform (TAP), which provides a developer experience on Kubernetes across multiple clouds. TAP aims to unlock developer productivity, build rapid paths to production, and coordinate the work of development, security and operations teams. It offers features like pre-configured templates, integrated developer tools, centralized visibility and workload status, role-based access control, automated pipelines and built-in security. The presentation provides examples of how these capabilities improve experiences for developers, operations teams and security teams.
The document provides information about a Tanzu Developer Connect Workshop on Tanzu Application Platform. The agenda includes welcome and introductions on Tanzu Application Platform, followed by interactive hands-on workshops on the developer experience and operator experience. It will conclude with a quiz, prizes and giveaways. The document discusses challenges with developing on Kubernetes and how Tanzu Application Platform aims to improve the developer experience with features like pre-configured templates, developer tools integration, rapid iteration and centralized management.
The Tanzu Developer Connect is a hands-on workshop that dives deep into TAP. Attendees receive a hands on experience. This is a great program to leverage accounts with current TAP opportunities.
The Tanzu Developer Connect is a hands-on workshop that dives deep into TAP. Attendees receive a hands on experience. This is a great program to leverage accounts with current TAP opportunities.
Simplify and Scale Enterprise Apps in the Cloud | Dallas 2023VMware Tanzu
This document discusses simplifying and scaling enterprise Spring applications in the cloud. It provides an overview of Azure Spring Apps, which is a fully managed platform for running Spring applications on Azure. Azure Spring Apps handles infrastructure management and application lifecycle management, allowing developers to focus on code. It is jointly built, operated, and supported by Microsoft and VMware. The document demonstrates how to create an Azure Spring Apps service, create an application, and deploy code to the application using three simple commands. It also discusses features of Azure Spring Apps Enterprise, which includes additional capabilities from VMware Tanzu components.
SpringOne Tour: Deliver 15-Factor Applications on Kubernetes with Spring BootVMware Tanzu
The document discusses 15 factors for building cloud native applications with Kubernetes based on the 12 factor app methodology. It covers factors such as treating code as immutable, externalizing configuration, building stateless and disposable processes, implementing authentication and authorization securely, and monitoring applications like space probes. The presentation aims to provide an overview of the 15 factors and demonstrate how to build cloud native applications using Kubernetes based on these principles.
SpringOne Tour: The Influential Software EngineerVMware Tanzu
The document discusses the importance of culture in software projects and how to influence culture. It notes that software projects involve people and personalities, not just technology. It emphasizes that culture informs everything a company does and is very difficult to change. It provides advice on being aware of your company's culture, finding ways to inculcate good cultural values like writing high-quality code, and approaches for influencing decision makers to prioritize culture.
SpringOne Tour: Domain-Driven Design: Theory vs PracticeVMware Tanzu
This document discusses domain-driven design, clean architecture, bounded contexts, and various modeling concepts. It provides examples of an e-scooter reservation system to illustrate domain modeling techniques. Key topics covered include identifying aggregates, bounded contexts, ensuring single sources of truth, avoiding anemic domain models, and focusing on observable domain behaviors rather than implementation details.
Automation Dreamin' 2022: Sharing Some Gratitude with Your UsersLynda Kane
Slide Deck from Automation Dreamin'2022 presentation Sharing Some Gratitude with Your Users on creating a Flow to present a random statement of Gratitude to a User in Salesforce.
Dev Dives: Automate and orchestrate your processes with UiPath MaestroUiPathCommunity
This session is designed to equip developers with the skills needed to build mission-critical, end-to-end processes that seamlessly orchestrate agents, people, and robots.
📕 Here's what you can expect:
- Modeling: Build end-to-end processes using BPMN.
- Implementing: Integrate agentic tasks, RPA, APIs, and advanced decisioning into processes.
- Operating: Control process instances with rewind, replay, pause, and stop functions.
- Monitoring: Use dashboards and embedded analytics for real-time insights into process instances.
This webinar is a must-attend for developers looking to enhance their agentic automation skills and orchestrate robust, mission-critical processes.
👨🏫 Speaker:
Andrei Vintila, Principal Product Manager @UiPath
This session streamed live on April 29, 2025, 16:00 CET.
Check out all our upcoming Dev Dives sessions at https://community.uipath.com/dev-dives-automation-developer-2025/.
Mobile App Development Company in Saudi ArabiaSteve Jonas
EmizenTech is a globally recognized software development company, proudly serving businesses since 2013. With over 11+ years of industry experience and a team of 200+ skilled professionals, we have successfully delivered 1200+ projects across various sectors. As a leading Mobile App Development Company In Saudi Arabia we offer end-to-end solutions for iOS, Android, and cross-platform applications. Our apps are known for their user-friendly interfaces, scalability, high performance, and strong security features. We tailor each mobile application to meet the unique needs of different industries, ensuring a seamless user experience. EmizenTech is committed to turning your vision into a powerful digital product that drives growth, innovation, and long-term success in the competitive mobile landscape of Saudi Arabia.
Leading AI Innovation As A Product Manager - Michael JidaelMichael Jidael
Unlike traditional product management, AI product leadership requires new mental models, collaborative approaches, and new measurement frameworks. This presentation breaks down how Product Managers can successfully lead AI Innovation in today's rapidly evolving technology landscape. Drawing from practical experience and industry best practices, I shared frameworks, approaches, and mindset shifts essential for product leaders navigating the unique challenges of AI product development.
In this deck, you'll discover:
- What AI leadership means for product managers
- The fundamental paradigm shift required for AI product development.
- A framework for identifying high-value AI opportunities for your products.
- How to transition from user stories to AI learning loops and hypothesis-driven development.
- The essential AI product management framework for defining, developing, and deploying intelligence.
- Technical and business metrics that matter in AI product development.
- Strategies for effective collaboration with data science and engineering teams.
- Framework for handling AI's probabilistic nature and setting stakeholder expectations.
- A real-world case study demonstrating these principles in action.
- Practical next steps to begin your AI product leadership journey.
This presentation is essential for Product Managers, aspiring PMs, product leaders, innovators, and anyone interested in understanding how to successfully build and manage AI-powered products from idea to impact. The key takeaway is that leading AI products is about creating capabilities (intelligence) that continuously improve and deliver increasing value over time.
AI and Data Privacy in 2025: Global TrendsInData Labs
In this infographic, we explore how businesses can implement effective governance frameworks to address AI data privacy. Understanding it is crucial for developing effective strategies that ensure compliance, safeguard customer trust, and leverage AI responsibly. Equip yourself with insights that can drive informed decision-making and position your organization for success in the future of data privacy.
This infographic contains:
-AI and data privacy: Key findings
-Statistics on AI data privacy in the today’s world
-Tips on how to overcome data privacy challenges
-Benefits of AI data security investments.
Keep up-to-date on how AI is reshaping privacy standards and what this entails for both individuals and organizations.
Special Meetup Edition - TDX Bengaluru Meetup #52.pptxshyamraj55
We’re bringing the TDX energy to our community with 2 power-packed sessions:
🛠️ Workshop: MuleSoft for Agentforce
Explore the new version of our hands-on workshop featuring the latest Topic Center and API Catalog updates.
📄 Talk: Power Up Document Processing
Dive into smart automation with MuleSoft IDP, NLP, and Einstein AI for intelligent document workflows.
Automation Hour 1/28/2022: Capture User Feedback from AnywhereLynda Kane
Slide Deck from Automation Hour 1/28/2022 presentation Capture User Feedback from Anywhere presenting setting up a Custom Object and Flow to collection User Feedback in Dynamic Pages and schedule a report to act on that feedback regularly.
Procurement Insights Cost To Value Guide.pptxJon Hansen
Procurement Insights integrated Historic Procurement Industry Archives, serves as a powerful complement — not a competitor — to other procurement industry firms. It fills critical gaps in depth, agility, and contextual insight that most traditional analyst and association models overlook.
Learn more about this value- driven proprietary service offering here.
Learn the Basics of Agile Development: Your Step-by-Step GuideMarcel David
New to Agile? This step-by-step guide is your perfect starting point. "Learn the Basics of Agile Development" simplifies complex concepts, providing you with a clear understanding of how Agile can improve software development and project management. Discover the benefits of iterative work, team collaboration, and flexible planning.
Hands On: Create a Lightning Aura Component with force:RecordDataLynda Kane
Slide Deck from the 3/26/2020 virtual meeting of the Cleveland Developer Group presentation on creating a Lightning Aura Component using force:RecordData.
AI Changes Everything – Talk at Cardiff Metropolitan University, 29th April 2...Alan Dix
Talk at the final event of Data Fusion Dynamics: A Collaborative UK-Saudi Initiative in Cybersecurity and Artificial Intelligence funded by the British Council UK-Saudi Challenge Fund 2024, Cardiff Metropolitan University, 29th April 2025
https://alandix.com/academic/talks/CMet2025-AI-Changes-Everything/
Is AI just another technology, or does it fundamentally change the way we live and think?
Every technology has a direct impact with micro-ethical consequences, some good, some bad. However more profound are the ways in which some technologies reshape the very fabric of society with macro-ethical impacts. The invention of the stirrup revolutionised mounted combat, but as a side effect gave rise to the feudal system, which still shapes politics today. The internal combustion engine offers personal freedom and creates pollution, but has also transformed the nature of urban planning and international trade. When we look at AI the micro-ethical issues, such as bias, are most obvious, but the macro-ethical challenges may be greater.
At a micro-ethical level AI has the potential to deepen social, ethnic and gender bias, issues I have warned about since the early 1990s! It is also being used increasingly on the battlefield. However, it also offers amazing opportunities in health and educations, as the recent Nobel prizes for the developers of AlphaFold illustrate. More radically, the need to encode ethics acts as a mirror to surface essential ethical problems and conflicts.
At the macro-ethical level, by the early 2000s digital technology had already begun to undermine sovereignty (e.g. gambling), market economics (through network effects and emergent monopolies), and the very meaning of money. Modern AI is the child of big data, big computation and ultimately big business, intensifying the inherent tendency of digital technology to concentrate power. AI is already unravelling the fundamentals of the social, political and economic world around us, but this is a world that needs radical reimagining to overcome the global environmental and human challenges that confront us. Our challenge is whether to let the threads fall as they may, or to use them to weave a better future.
#3: Our Roots – Why you should have confidence in YB
#4: Founded by a team from Facebook
9 members of the core Data Infrastructure team @ FB
From 2006-2013 Unique journey …Started off on Bare-Metal….moved to Containers…..had to address multiple DC’s in a very short time……over 1 Billion people needing low latency reads….all across the planet
FaceBook Messenger - Inbox/Messages
Operations Data Store
Site Integrity Application
Fraud Detections
Needed strong consistency for Site Integrity and Fraud so they created H-Base
Determined there was a strong need for a Cloud-Based DB Platform
Maturity of the company…..went GA with 1.0 in April
Added Oracle , Nutanics personnel
Just closed a round with LightSpeed and Dell Technologies Capital
Brought 1.0 to market April 2018
Scaled from 30M to 1.2 B
#9: Custom tier abstracts the complexity…
Transactional
Performant
Scales
Devloper agility
Open API’s you already know/use
Our API’s extend the capabilities….our SQL functionality is now in NoSQL….and NoSQL functionality is now in SQL
#10: Does this look like your environment?
Describe pain/process for the developers
How do they sync ----pushing it up into the app layer….you end up implementing work-arounds….slowing your dev cycles
Data tier is usually behind the application teir
Application tier is usually stateless
Transactional data written in int an SQL DB that is manually sharded…..CLICK,….now you need to be in 2 data centers so you have to replicate the data…So you usually have Cassandra or MongoDB….data starts to get silo’ed in the organizations…and because performance isnt what it should be….you introduce a caching layer…..such as Redis
So….you end up with 4 data stores…..a lot of complexity…..and your architecture becomes BRITTLE…
Does moving to the cloud affect this…..NO…..you are just using cloud-versions on these solutions….
Cassandra Consistency
R+W>N (N=replication factor) Quorum is easy when you are using eventual consistency
Consistency – Strong vs Eventual
3 reasons why this is not agile
3 observations why it is operationally complex
#12: We built a new core
New core engine for the data fabric
Open Standards
Purpose built data fabric – a new cloud-native database not another legacy database
Multi-cloud
Nomenclature check
#13: Cassandra friendly
Existing Cassandra apps can run against YB
#15: Turvo
Built app that unifies all the functions they need to monitor
Started with Mongo
Narvar
OEM for doing customer experience…..
Handles the entire customer experience
Started on Dynamo and Elasticache
SQL and NoSQL requirements meant we were good fit for them
For Retailers….very important to scale to meet holiday buying season
#27: Multi-cloud set up
Zero down-time migrations and upgrade
Rolling upgrades
Trivial to bring on new regions
Trivial to bring on new IaaS
- Including database
#29: This use case is Good to orient you around how YugaByte provides value for an organization
This is a web conferencing application (Video/audio/chat)
The service using YugaByte is for user-indentity: User name/password/attributes
Millions of users
Individuals
Corporations
Important to know where they were loggin in from
Build out slide....
User name needs to be strongly consistent….can’t be eventually……
Initial write (username & Password) is written in NoSQL DB….(MySQL) ….which is replicated for redundancy in a master/slave configuration…..all of the writes were happening in the US….but the reads where happening across the globe so they replicated the data in Cassandra and used that to stage the data across the globe….and to achieve low latency reads….they were using CouchBase as caching tier
Lots of complexity…..which leads to cost….but the biggest cost was to opportunity-cost. When ever there was a change to the application, the developers needed to coordinate across 3 different DB’s
With Yugabyte….you have one cluster taking rights (of an homogenous DB)…and then data replicated to 3 data centers to achieve low-latency reads.
BENEFITS:
Writing to 1 platform
We maintain data resiliency
Feature velocity / Agility for the company
#31: Focuses on Redis as a DB
Redis – great as a caching tier….benefits of ease of use for some data types
Large website for obtaining news, stock tickers and information about companies …..and to persist the data it was written to a homegrown DB.
Far right side…..Large Redis cluster was used to achieve low latency reads. Adding to functionality/applications was very time consuming…..it meant they had to manually Shard the cluster….which made it VERY brittle (usually took 6-9 months)…started a process of…..deploy additional nodes…..reshard…re-deploy
If you look….you had to write to 2 separate DB’s….which creates additional complexity
#32: Now…with Yugabyte….
They write into 1 cluster (using the Redis API)
Reads come into the Redis tier from the app itself…low latency reads are served up….single Milisecond (6 or less)
Started with 4 node cluster
Doubled to 8 Nodes only took 30 minutes (versus 6 – 9 months)
They also wanted the ability to birst into the cloud….(which we do via our read replication)
Multi-Cloud…Multi-Datacenter
#35: Turvo
Built app that unifies all the functions they need to monitor
Started with Mongo
Narvar
OEM for doing customer experience…..
Handles the entire customer experience
Started on Dynamo and Elasticache
SQL and NoSQL requirements meant we were good fit for them
For Retailers….very important to scale to meet holiday buying season