The document outlines an agenda for a MongoDB event in Frankfurt on November 30th 2017. The agenda includes introductions, implementing a cloud-based data strategy, best practices for migrating from RDBMS to MongoDB, how MongoDB can provide support, and a Q&A session. It also lists the speakers which include representatives from MongoDB and Bosch Software Innovations.
Presented by Rob Walters, Solutions Architect, MongoDB, at MongoDB Evenings New England 2017.
MongoDB 3.6 is the latest version of the world's most popular document database. In this session we will cover the key themes of the release including speed to develop, speed to production and speed to insight. Learn about the key features that support these themes and how you can start leveraging them today!
MongoDB is a leading database technology that combines the foundations of RDBMS with the innovations of NoSQL, allowing organizations to simultaneously boost productivity and lower TCO.
MongoDB Enterprise Advanced is a finely-tuned package of advanced software, enterprise-grade support, and other services designed to accelerate your success with MongoDB in every stage of your app lifecycle, from early development to the scale-out of mission-critical production environments.
With the release of 3.2, MongoDB Enterprise Advanced now includes:
MongoDB Ops Manager 2.0
MongoDB Compass, the MongoDB GUI
MongoDB Connector for Business Intelligence
Encrypted Storage Engine
In-Memory Storage Engine (beta)
Attend this webinar to learn how MongoDB Enterprise Advanced can help you get to market faster and de-risk your mission critical deployments.
MongoDB 3.6 helps you *move at the speed of your data* - turning developers, operations teams, and analysts into a growth engine for the business. It enables new apps to be delivered to market faster, running reliably and securely at scale, and unlocking insights and intelligence in real time. Learn more: https://www.mongodb.com/mongodb-3.6
Webinar: Simplifying the Database Experience with MongoDB AtlasMongoDB
MongoDB Atlas is our database as a service for MongoDB. In this webinar you’ll learn how it provides all of the features of MongoDB, without all of the operational heavy lifting, and all through a pay-as-you-go model billed on an hourly basis.
Presented by Radu Craioveanu, Director of Software Development, Clinical Systems, Fresenius Medical Care at MongoDB Evenings New England 2017.
resenius is a large healthcare enterprise, specializing in dialysis care. Fresenius' 40,000 clinicians and physicians deliver over 100,000 dialysis treatments per day, across 8 time zones. There is significant pressure to improve treatment outcomes, lower costs, expand the patient coverage, and overall become a Value Based Services provider, sharing the risk with the payers, the insurance companies. This pressures requires Fresenius to adapt, change, and leverage technologies and processes that can enable a rapid transformation to Value Based Care. </br></br>Using technologies and partnerships with players such as MongoDB, Red Hat, and others with a similar, open source innovative approach to progress, Fresenius has been able to implement a healthcare platform that is the foundation onto which the business can transform itself.
MongoDB has enabled Fresenius to achieve high availability of systems across multiple data centers, a data lake concept used for predictive analytics and reporting, enhanced messaging capabilities, fast, effective and distributed archiving, rapid application development via MEAN stacks, and Red Hat Open Shift Docker Container ready-to-use persistence.
Which Change Data Capture Strategy is Right for You?Precisely
Change Data Capture or CDC is the practice of moving the changes made in an important transactional system to other systems, so that data is kept current and consistent across the enterprise. CDC keeps reporting and analytic systems working on the latest, most accurate data.
Many different CDC strategies exist. Each strategy has advantages and disadvantages. Some put an undue burden on the source database. They can cause queries or applications to become slow or even fail. Some bog down network bandwidth, or have big delays between change and replication.
Each business process has different requirements, as well. For some business needs, a replication delay of more than a second is too long. For others, a delay of less than 24 hours is excellent.
Which CDC strategy will match your business needs? How do you choose?
View this webcast on-demand to learn:
• Advantages and disadvantages of different CDC methods
• The replication latency your project requires
• How to keep data current in Big Data technologies like Hadoop
In-memory computing stores information in RAM rather than on disk drives for faster access. It allows companies to analyze large amounts of data quickly and perform operations more efficiently. As memory prices drop, in-memory computing is becoming more widespread. Some companies like SAP and Oracle have adapted in-memory concepts, processing data 1000 times faster. In-memory databases provide advantages like faster transactions and high stability for applications requiring quick response times.
During this presentation, Infusion and MongoDB shared their mainframe optimization experiences and best practices. These have been gained from working with a variety of organizations, including a case study from one of the world’s largest banks. MongoDB and Infusion bring a tested approach that provides a new way of modernizing mainframe applications, while keeping pace with the demand for new digital services.
<b>Elevate MongoDB with ODBC/JDBC </b>[4:05 pm - 4:25 pm]<br />Adoption for MongoDB is growing across the enterprise and disrupting existing business intelligence, analytics and data integration infrastructure. Join us to disrupt that disruption using ODBC and JDBC access to MongoDB for instant out-of-box integration with existing infrastructure to elevate and expand your organization’s MongoDB footprint. We'll talk about common challenges and gotchas that shops face when exposing unstructured and semi-structured data using these established data connectivity standards. Existing infrastructure requirements should not dictate developers’ freedom of choice in a database
NoSQL and Spatial Database Capabilities using PostgreSQLEDB
PostgreSQL is an object-relational database system. NoSQL on the other hand is a non-relational database and is document-oriented. Learn how the PostgreSQL database gives one the flexible options to combine NoSQL workloads with the relational query power by offering JSON data types. With PostgreSQL, new capabilities can be developed and plugged into the database as required.
Attend this webinar to learn:
- The new features and capabilities in PostgreSQL for new workloads, requiring greater flexibility in the data model
- NoSQL with JSON, Hstore and its performance and features for enterprises
- Spatial SQL - advanced features in PostGIS application with PostGIS extension
RedisConf18 - Scaling Whitepages With Redison FlashRedis Labs
This document discusses Whitepages' transition from self-hosted Redis to Redis on Flash (Redis Labs) to store its Global Identity Graph. Some key points:
- Whitepages stores a large identity graph with 5 billion entities and 18 billion links used for people search and identity verification.
- Previous solutions like Cassandra, MongoDB, and self-hosted Redis had latency or cost issues at Whitepages' scale.
- Redis on Flash provided RAM-like latencies using flash storage at a lower cost - reducing required nodes by 40% and costs by 60% while maintaining performance.
- The transition was easy as Redis on Flash could be put into production immediately without changes to existing infrastructure or requiring extra training.
Caching for Microservices Architectures: Session II - Caching PatternsVMware Tanzu
In the first webinar of the series we covered the importance of caching in microservice-based application architectures—in addition to improving performance it also aids in making content available from legacy systems, promotes loose coupling and team autonomy, and provides air gaps that can limit failures from cascading through a system.
To reap these benefits, though, the right caching patterns must be employed. In this webinar, we will examine various caching patterns and shed light on how they deliver the capabilities needed by our microservices. What about rapidly changing data, and concurrent updates to data? What impact do these and other factors have to various use cases and patterns?
Understanding data access patterns, covered in this webinar, will help you make the right decisions for each use case. Beyond the simplest of use cases, caching can be tricky business—join us for this webinar to see how best to use them.
Jagdish Mirani, Cornelia Davis, Michael Stolz, Pulkit Chandra, Pivotal
Organisations are building their applications around microservice architectures because of the flexibility, speed of delivery, and maintainability they deliver. In this session, the concepts behind microservices, containers and orchestration was explained and how to use them with MongoDB.
Webinar: Data Streaming with Apache Kafka & MongoDBMongoDB
A new generation of technologies is needed to consume and exploit today's real time, fast moving data sources. Apache Kafka, originally developed at LinkedIn, has emerged as one of these key new technologies.
Mainframe Modernization with Precisely and Microsoft AzurePrecisely
Today’s businesses are leveraging Microsoft Azure to modernize operations, transform customer experience, and increase profit. However, if the rich data generated by the mainframe applications is missed in the move to the cloud, you miss the mark.
Without the right solutions in place, migrating mainframe data to Microsoft Azure is expensive, time-consuming, and reliant on highly specialized skillsets. Precisely Connect can quickly integrate mainframe data at scale into Microsoft Azure without sacrificing functionality, security, or ease of use.
View this on-demand webinar to hear from Microsoft Azure and Precisely data integration experts. You will:
- Learn how to build highly scalable, reliable data pipelines between the mainframe and Microsoft Azure services
- Understand how to make your Microsoft Azure implementation ready for mainframe
- Dive into case studies of businesses that have successfully included mainframe data in their cloud modernization efforts with Precisely and Microsoft Azure
Delivering fast, powerful and scalable analyticsMariaDB plc
This document discusses different types of analytics and provides examples of how MariaDB AX can be used for fast and scalable analytics. It describes descriptive, diagnostic, predictive, and prescriptive analytics. It then gives examples of how MariaDB AX has been used by organizations like IHME for large-scale health data analytics, by CIM for population health analysis, and by Genus for genetic profiling of livestock. MariaDB AX provides high-performance analytics through features like columnar storage and parallel query processing.
Making the Most of Data in Multiple Data Sources (with Virtual Data Lakes)DataWorks Summit
Most organizations today implement different data stores to support business operations. As a result, data ends up stored across a multitude of often heterogenous systems, like RDBMS, NoSQL, data warehouses, data marts, Hadoop, etc., with limited interaction and/or interoperability between them. The end result is often a vast eco-system of data stores with different "temperature" data, some level of duplication and, no effective way of bringing it all together for business analytics. With such disparate data, how can an organization exploit the wealth of information? This opens up the need for proven techniques to quickly and easily deliver the data to the people who need it. In this session, you'll see how to modernize your enterprise by making data accessible with enterprise capabilities like querying using SQL, granular security for data access, and maintaining high query performance and high concurrency.
Big Data Business Transformation - Big Picture and BlueprintsAshnikbiz
Kaustubh Patwardhan, Head of Strategy and Business Development at Ashnik presents the big picture and blueprints of a big data journey for enterprises. The Value of Big Data – Machine Learning and its big impact. He covers a spectrum of Big Data use cases where right data storage, integration & data consolidation plays a big role.
This document provides an overview of MariaDB's 2017 roadshow, including what they are doing, where they are going, and who the field CTO is. It discusses trends in the database market moving away from expensive proprietary databases toward lower-cost open source options with subscriptions and community involvement. It highlights cost savings of MariaDB compared to Oracle and MariaDB's extensible architecture and community contributions. It also summarizes MariaDB products and technologies like the database server, MaxScale proxy, and ColumnStore, as well as MariaDB's customers, use cases, services, and how to get started with MariaDB.
LinkedIn Infrastructure (analytics@webscale, at fb 2013)Jun Rao
This is the presentation at analytics@webscale in 2013 (http://analyticswebscale.splashthat.com/?em=187&utm_campaign=website&utm_source=sg&utm_medium=em)
Speaker: Jerry Reghunadh, Architect, CAPIOT Software Pvt. Ltd.
Level: 200 (Intermediate)
Track: Microservices
One of the leading assisted e-commerce players in India approached CAPIOT to rebuild their ERP system from the ground up. Their existing PHP-MySQL setup, while rich in functionality and having served them well for under half a decade, would not scale to meet future demands due to the exponential grown they were experiencing.
We built the entire system using a microservices architecture. To develop APIs we used Node.js, Express, Swagger and Mongoose, and MongoDB was used as the active data store. During the development phase, we solved several problems ranging from cross-service calls, data consistency, service discovery, and security.
One of the issues that we faced is how to effectively design and make cross-service calls. Should we implement a cross-service call for every document that we require or should we duplicate and distribute the data, reducing cross-service calls? We found a balance between these two and engineered a solution that gave us good performance.
In addition, our current system has 36 independent services. We enabled services to auto-discover and make secure calls.
We used Swagger to define our APIs first and enforce request and response validations and Mongoose as our ODM for schema validation. We also heavily depend on pre-save hooks to validate data and post-save hooks to trigger changes in other systems. This API-driven approach vastly enabled our frontend and backend teams to scrum together on a single API spec without worrying about the repercussions of changing API schemas.
What You Will Learn:
- How we used Swagger and Mongoose to off-load validations and schema enforcements. We used Swagger to define our APIs first and enforce request and response validations and Mongoose as our ODM for schema validation. We also heavily depend on pre-save hooks to validate data and post-save hooks to trigger changes in other systems. This API-driven approach vastly enabled our frontend and backend teams to scrum together on a single API spec without worrying about the repercussions of changing API schemas.
- How microservices and cross-service calls work. One of the issues that we faced is how to effectively design and make cross-service calls. Should we implement a cross-service call for every document that we require or should we duplicate and distribute the data, reducing cross-service calls? We found a balance between these two and engineered a solution that gave us good performance.
- How we implemented microservice auto discovery: Our current system has 36 independent services, so we enabled services to auto-discover and make secure calls.
Apache frameworks provide solutions for processing big and fast data. Traditional APIs use a request/response model with pull-based interactions, while modern data streaming uses a publish/subscribe model. Key concepts for big data architectures include batch processing frameworks like Hadoop, stream processing tools like Storm, and hybrid options like Spark and Flink. Popular data ingestion tools include Kafka for messaging, Flume for log data, and Sqoop for structured data. The best solution depends on requirements like latency, data volume, and workload type.
Qlik and Confluent Success Stories with Kafka - How Generali and Skechers Kee...HostedbyConfluent
Converting production databases into live data streams for Apache Kafka can be labor intensive and costly. As Kafka architectures grow, complexity also rises as data teams begin to configure clusters for redundancy, partitions for performance, as well as for consumer groups for correlated analytics processing. In this breakout session, you’ll hear data streaming success stories from Generali and Skechers that leverage Qlik Data Integration and Confluent. You’ll discover how Qlik’s data integration platform lets organizations automatically produce real-time transaction streams into Kafka, Confluent Platform, or Confluent Cloud, deliver faster business insights from data, enable streaming analytics, as well as streaming ingestion for modern analytics. Learn how these customer use Qlik and Confluent to: - Turn databases into live data feeds - Simplify and automate the real-time data streaming process - Accelerate data delivery to enable real-time analytics Learn how Skechers and Generali breathe new life into data in the cloud, stay ahead of changing demands, while lowering over-reliance on resources, production time and costs.
Migrating on premises workload to azure sql databasePARIKSHIT SAVJANI
This document provides an overview of migrating databases from on-premises SQL Server to Azure SQL Database Managed Instance. It discusses why companies are moving to the cloud, challenges with migration, and the tools and services available to help with assessment and migration including Data Migration Service. Key steps in the migration workflow include assessing the database and application, addressing compatibility issues, and deploying the converted schema to Managed Instance which provides high compatibility with on-premises SQL Server in a fully managed platform as a service model.
This document compares the total cost of ownership of MongoDB and Oracle databases. It outlines the various cost categories to consider, including upfront costs like software, hardware, development efforts, and ongoing costs like maintenance and support. The document then provides two example scenarios - a smaller and larger enterprise project - comparing the expected costs of building each using MongoDB versus Oracle. It finds that for these examples, using MongoDB is over 70% less expensive than using Oracle. Finally, it discusses how MongoDB's advantages in flexibility, ease of use and support for modern development can help reduce costs and speed development.
De nouvelles générations de technologies de bases de données permettent aux organisations de créer des applications jusque-là inédites, à une vitesse et une échelle inimaginables auparavant. MongoDB est la base de données qui connaît la croissance la plus rapide au monde. La nouvelle version 3.2 offre les avantages des architectures de bases de données modernes à une gamme toujours plus large d'applications et d'utilisateurs.
MongoDB es la base de datos con más rápido crecimiento del mundo La nueva versión 3.2 extiende los beneficios de las modernas arquitecturas de bases de datos a una gama aun más amplia de aplicaciones y usuarios.
En esta grabación del seminario web presentamos todas las novedades, que incluyen:
● Nuevos motores de almacenamiento conectables.
● Una visión empresarial más rápida con búsquedas y análisis mejorados en tiempo real, combinada con una conectividad fluida a herramientas de BI estándar.
● Gestión de datos simplificado con validación de documentos, junto a una detección y visualización de esquema basadas en una interfaz gráfica.
Mayor eficacia operativa con plataformas de gestión mejoradas, disponibilidad continua en implementaciones multirregionales y distribuidas, y actualizaciones con inactividad cero.
During this presentation, Infusion and MongoDB shared their mainframe optimization experiences and best practices. These have been gained from working with a variety of organizations, including a case study from one of the world’s largest banks. MongoDB and Infusion bring a tested approach that provides a new way of modernizing mainframe applications, while keeping pace with the demand for new digital services.
<b>Elevate MongoDB with ODBC/JDBC </b>[4:05 pm - 4:25 pm]<br />Adoption for MongoDB is growing across the enterprise and disrupting existing business intelligence, analytics and data integration infrastructure. Join us to disrupt that disruption using ODBC and JDBC access to MongoDB for instant out-of-box integration with existing infrastructure to elevate and expand your organization’s MongoDB footprint. We'll talk about common challenges and gotchas that shops face when exposing unstructured and semi-structured data using these established data connectivity standards. Existing infrastructure requirements should not dictate developers’ freedom of choice in a database
NoSQL and Spatial Database Capabilities using PostgreSQLEDB
PostgreSQL is an object-relational database system. NoSQL on the other hand is a non-relational database and is document-oriented. Learn how the PostgreSQL database gives one the flexible options to combine NoSQL workloads with the relational query power by offering JSON data types. With PostgreSQL, new capabilities can be developed and plugged into the database as required.
Attend this webinar to learn:
- The new features and capabilities in PostgreSQL for new workloads, requiring greater flexibility in the data model
- NoSQL with JSON, Hstore and its performance and features for enterprises
- Spatial SQL - advanced features in PostGIS application with PostGIS extension
RedisConf18 - Scaling Whitepages With Redison FlashRedis Labs
This document discusses Whitepages' transition from self-hosted Redis to Redis on Flash (Redis Labs) to store its Global Identity Graph. Some key points:
- Whitepages stores a large identity graph with 5 billion entities and 18 billion links used for people search and identity verification.
- Previous solutions like Cassandra, MongoDB, and self-hosted Redis had latency or cost issues at Whitepages' scale.
- Redis on Flash provided RAM-like latencies using flash storage at a lower cost - reducing required nodes by 40% and costs by 60% while maintaining performance.
- The transition was easy as Redis on Flash could be put into production immediately without changes to existing infrastructure or requiring extra training.
Caching for Microservices Architectures: Session II - Caching PatternsVMware Tanzu
In the first webinar of the series we covered the importance of caching in microservice-based application architectures—in addition to improving performance it also aids in making content available from legacy systems, promotes loose coupling and team autonomy, and provides air gaps that can limit failures from cascading through a system.
To reap these benefits, though, the right caching patterns must be employed. In this webinar, we will examine various caching patterns and shed light on how they deliver the capabilities needed by our microservices. What about rapidly changing data, and concurrent updates to data? What impact do these and other factors have to various use cases and patterns?
Understanding data access patterns, covered in this webinar, will help you make the right decisions for each use case. Beyond the simplest of use cases, caching can be tricky business—join us for this webinar to see how best to use them.
Jagdish Mirani, Cornelia Davis, Michael Stolz, Pulkit Chandra, Pivotal
Organisations are building their applications around microservice architectures because of the flexibility, speed of delivery, and maintainability they deliver. In this session, the concepts behind microservices, containers and orchestration was explained and how to use them with MongoDB.
Webinar: Data Streaming with Apache Kafka & MongoDBMongoDB
A new generation of technologies is needed to consume and exploit today's real time, fast moving data sources. Apache Kafka, originally developed at LinkedIn, has emerged as one of these key new technologies.
Mainframe Modernization with Precisely and Microsoft AzurePrecisely
Today’s businesses are leveraging Microsoft Azure to modernize operations, transform customer experience, and increase profit. However, if the rich data generated by the mainframe applications is missed in the move to the cloud, you miss the mark.
Without the right solutions in place, migrating mainframe data to Microsoft Azure is expensive, time-consuming, and reliant on highly specialized skillsets. Precisely Connect can quickly integrate mainframe data at scale into Microsoft Azure without sacrificing functionality, security, or ease of use.
View this on-demand webinar to hear from Microsoft Azure and Precisely data integration experts. You will:
- Learn how to build highly scalable, reliable data pipelines between the mainframe and Microsoft Azure services
- Understand how to make your Microsoft Azure implementation ready for mainframe
- Dive into case studies of businesses that have successfully included mainframe data in their cloud modernization efforts with Precisely and Microsoft Azure
Delivering fast, powerful and scalable analyticsMariaDB plc
This document discusses different types of analytics and provides examples of how MariaDB AX can be used for fast and scalable analytics. It describes descriptive, diagnostic, predictive, and prescriptive analytics. It then gives examples of how MariaDB AX has been used by organizations like IHME for large-scale health data analytics, by CIM for population health analysis, and by Genus for genetic profiling of livestock. MariaDB AX provides high-performance analytics through features like columnar storage and parallel query processing.
Making the Most of Data in Multiple Data Sources (with Virtual Data Lakes)DataWorks Summit
Most organizations today implement different data stores to support business operations. As a result, data ends up stored across a multitude of often heterogenous systems, like RDBMS, NoSQL, data warehouses, data marts, Hadoop, etc., with limited interaction and/or interoperability between them. The end result is often a vast eco-system of data stores with different "temperature" data, some level of duplication and, no effective way of bringing it all together for business analytics. With such disparate data, how can an organization exploit the wealth of information? This opens up the need for proven techniques to quickly and easily deliver the data to the people who need it. In this session, you'll see how to modernize your enterprise by making data accessible with enterprise capabilities like querying using SQL, granular security for data access, and maintaining high query performance and high concurrency.
Big Data Business Transformation - Big Picture and BlueprintsAshnikbiz
Kaustubh Patwardhan, Head of Strategy and Business Development at Ashnik presents the big picture and blueprints of a big data journey for enterprises. The Value of Big Data – Machine Learning and its big impact. He covers a spectrum of Big Data use cases where right data storage, integration & data consolidation plays a big role.
This document provides an overview of MariaDB's 2017 roadshow, including what they are doing, where they are going, and who the field CTO is. It discusses trends in the database market moving away from expensive proprietary databases toward lower-cost open source options with subscriptions and community involvement. It highlights cost savings of MariaDB compared to Oracle and MariaDB's extensible architecture and community contributions. It also summarizes MariaDB products and technologies like the database server, MaxScale proxy, and ColumnStore, as well as MariaDB's customers, use cases, services, and how to get started with MariaDB.
LinkedIn Infrastructure (analytics@webscale, at fb 2013)Jun Rao
This is the presentation at analytics@webscale in 2013 (http://analyticswebscale.splashthat.com/?em=187&utm_campaign=website&utm_source=sg&utm_medium=em)
Speaker: Jerry Reghunadh, Architect, CAPIOT Software Pvt. Ltd.
Level: 200 (Intermediate)
Track: Microservices
One of the leading assisted e-commerce players in India approached CAPIOT to rebuild their ERP system from the ground up. Their existing PHP-MySQL setup, while rich in functionality and having served them well for under half a decade, would not scale to meet future demands due to the exponential grown they were experiencing.
We built the entire system using a microservices architecture. To develop APIs we used Node.js, Express, Swagger and Mongoose, and MongoDB was used as the active data store. During the development phase, we solved several problems ranging from cross-service calls, data consistency, service discovery, and security.
One of the issues that we faced is how to effectively design and make cross-service calls. Should we implement a cross-service call for every document that we require or should we duplicate and distribute the data, reducing cross-service calls? We found a balance between these two and engineered a solution that gave us good performance.
In addition, our current system has 36 independent services. We enabled services to auto-discover and make secure calls.
We used Swagger to define our APIs first and enforce request and response validations and Mongoose as our ODM for schema validation. We also heavily depend on pre-save hooks to validate data and post-save hooks to trigger changes in other systems. This API-driven approach vastly enabled our frontend and backend teams to scrum together on a single API spec without worrying about the repercussions of changing API schemas.
What You Will Learn:
- How we used Swagger and Mongoose to off-load validations and schema enforcements. We used Swagger to define our APIs first and enforce request and response validations and Mongoose as our ODM for schema validation. We also heavily depend on pre-save hooks to validate data and post-save hooks to trigger changes in other systems. This API-driven approach vastly enabled our frontend and backend teams to scrum together on a single API spec without worrying about the repercussions of changing API schemas.
- How microservices and cross-service calls work. One of the issues that we faced is how to effectively design and make cross-service calls. Should we implement a cross-service call for every document that we require or should we duplicate and distribute the data, reducing cross-service calls? We found a balance between these two and engineered a solution that gave us good performance.
- How we implemented microservice auto discovery: Our current system has 36 independent services, so we enabled services to auto-discover and make secure calls.
Apache frameworks provide solutions for processing big and fast data. Traditional APIs use a request/response model with pull-based interactions, while modern data streaming uses a publish/subscribe model. Key concepts for big data architectures include batch processing frameworks like Hadoop, stream processing tools like Storm, and hybrid options like Spark and Flink. Popular data ingestion tools include Kafka for messaging, Flume for log data, and Sqoop for structured data. The best solution depends on requirements like latency, data volume, and workload type.
Qlik and Confluent Success Stories with Kafka - How Generali and Skechers Kee...HostedbyConfluent
Converting production databases into live data streams for Apache Kafka can be labor intensive and costly. As Kafka architectures grow, complexity also rises as data teams begin to configure clusters for redundancy, partitions for performance, as well as for consumer groups for correlated analytics processing. In this breakout session, you’ll hear data streaming success stories from Generali and Skechers that leverage Qlik Data Integration and Confluent. You’ll discover how Qlik’s data integration platform lets organizations automatically produce real-time transaction streams into Kafka, Confluent Platform, or Confluent Cloud, deliver faster business insights from data, enable streaming analytics, as well as streaming ingestion for modern analytics. Learn how these customer use Qlik and Confluent to: - Turn databases into live data feeds - Simplify and automate the real-time data streaming process - Accelerate data delivery to enable real-time analytics Learn how Skechers and Generali breathe new life into data in the cloud, stay ahead of changing demands, while lowering over-reliance on resources, production time and costs.
Migrating on premises workload to azure sql databasePARIKSHIT SAVJANI
This document provides an overview of migrating databases from on-premises SQL Server to Azure SQL Database Managed Instance. It discusses why companies are moving to the cloud, challenges with migration, and the tools and services available to help with assessment and migration including Data Migration Service. Key steps in the migration workflow include assessing the database and application, addressing compatibility issues, and deploying the converted schema to Managed Instance which provides high compatibility with on-premises SQL Server in a fully managed platform as a service model.
This document compares the total cost of ownership of MongoDB and Oracle databases. It outlines the various cost categories to consider, including upfront costs like software, hardware, development efforts, and ongoing costs like maintenance and support. The document then provides two example scenarios - a smaller and larger enterprise project - comparing the expected costs of building each using MongoDB versus Oracle. It finds that for these examples, using MongoDB is over 70% less expensive than using Oracle. Finally, it discusses how MongoDB's advantages in flexibility, ease of use and support for modern development can help reduce costs and speed development.
De nouvelles générations de technologies de bases de données permettent aux organisations de créer des applications jusque-là inédites, à une vitesse et une échelle inimaginables auparavant. MongoDB est la base de données qui connaît la croissance la plus rapide au monde. La nouvelle version 3.2 offre les avantages des architectures de bases de données modernes à une gamme toujours plus large d'applications et d'utilisateurs.
MongoDB es la base de datos con más rápido crecimiento del mundo La nueva versión 3.2 extiende los beneficios de las modernas arquitecturas de bases de datos a una gama aun más amplia de aplicaciones y usuarios.
En esta grabación del seminario web presentamos todas las novedades, que incluyen:
● Nuevos motores de almacenamiento conectables.
● Una visión empresarial más rápida con búsquedas y análisis mejorados en tiempo real, combinada con una conectividad fluida a herramientas de BI estándar.
● Gestión de datos simplificado con validación de documentos, junto a una detección y visualización de esquema basadas en una interfaz gráfica.
Mayor eficacia operativa con plataformas de gestión mejoradas, disponibilidad continua en implementaciones multirregionales y distribuidas, y actualizaciones con inactividad cero.
New generations of database technologies are allowing organizations to build applications never before possible, at a speed and scale that were previously unimaginable. MongoDB is the fastest growing database on the planet, and the new 3.2 release will bring the benefits of modern database architectures to an ever broader range of applications and users.
This presentation contains a preview of MongoDB 3.2 upcoming release where we explore the new storage engines, aggregation framework enhancements and utility features like document validation and partial indexes.
In-Memory Storage Engine (beta)
WiredTiger as the default storage engine
Advanced security (encryption at rest)
Document Validation
Advanced full text
Dynamic Lookups
BI Connector (Tableau, Qlikview, Cognos, BusinessObjects, etc...)
Database GUI with MongoDB Compass
And more...
MongoDB 3.2 introduces a host of new features and benefits, including encryption at rest, document validation, MongoDB Compass, numerous improvements to queries and the aggregation framework, and more. To take advantage of these features, your team needs an upgrade plan.
In this session, we’ll walk you through how to build an upgrade plan. We’ll show you how to validate your existing deployment, build a test environment with a representative workload, and detail how to carry out the upgrade. By the end, you should be prepared to start developing an upgrade plan for your deployment.
Technical feature review of features introduced by MongoDB 3.4 on graph capabilities, MongoDB UI tool: Compass, improvements on the replication and aggregation framework stages and utils. Operations improvements on Ops Manager and MongoDB Atlas.
Introduction to MongoDB and its best practicesAshishRathore72
This document provides a summary of a presentation on MongoDB best practices. It discusses MongoDB concepts like data modeling, CRUD operations, querying, and aggregation. It also covers topics like MongoDB security, scaling options, real-world use cases, and best practices for hardware, schema design, indexing, and scalability. The presentation provides an overview of using MongoDB effectively.
MongoDB 3.4 is a multi-model database that supports documents, relational data, key-value, and graph structures. It features new capabilities like faceted navigation for advanced analytics, views for mission critical applications, and intra-cluster compression. MongoDB also provides enterprise tools like Ops Manager for high resolution monitoring, Compass for visual data exploration, and connectors for BI and SQL tools.
Webinar: Faster Big Data Analytics with MongoDBMongoDB
Learn how to leverage MongoDB and Big Data technologies to derive rich business insight and build high performance business intelligence platforms. This presentation includes:
- Uncovering Opportunities with Big Data analytics
- Challenges of real-time data processing
- Best practices for performance optimization
- Real world case study
This presentation was given in partnership with CIGNEX Datamatics.
Webinar: Best Practices for Upgrading to MongoDB 3.0MongoDB
MongoDB 3.0 brings major enhancements. Write performance has improved by 7-10x with WiredTiger and document-level concurrency control. Compression reduces storage needs by up to 80%. To take advantage of these features, your team needs an upgrade plan.
In this session, we’ll walk you through how to build an upgrade plan. We’ll show you how to validate your existing deployment, build a test environment with a representative workload, and detail how to carry out the upgrade. You’ll walk away confident that you're prepared to upgrade.
This document discusses combining Apache Spark and MongoDB for real-time analytics. It describes how MongoDB provides rich analytics capabilities through queries, aggregations, and indexing. Apache Spark can further extend MongoDB's analytics by offering additional processing capabilities. Together, Spark and MongoDB enable organizations to perform real-time analytics directly on operational data without needing separate analytics infrastructure.
1. The document discusses adapting data strategies for the cloud, where time to market has replaced cost as the primary driver of cloud adoption.
2. It outlines key considerations for choosing a cloud data platform, including deployment flexibility, reducing complexity, agility, resiliency, scalability, cost, and security.
3. The document summarizes how MongoDB can provide a flexible cloud data strategy through offerings like MongoDB Atlas that offer deployment flexibility across public, private, and hybrid clouds without vendor lock-in.
La creación de una capa operacional con MongoDBMongoDB
The document discusses using MongoDB to modernize mainframe systems by reducing costs and increasing flexibility. It describes 5 phases of mainframe modernization with MongoDB, from initially offloading reads to using MongoDB as the primary system of record. Case studies are presented where MongoDB helped customers increase developer productivity by 5-10x, lower mainframe costs by 80%, and transform IT strategies by simplifying technology stacks.
MongoDB Europe 2016 - The Rise of the Data LakeMongoDB
The document discusses the rise of data lakes and how MongoDB can be used to build modern data management architectures. It provides examples of how companies like a Spanish bank and an insurance leader used MongoDB to create a single customer view across siloed data sources and improve customer experiences. The document also outlines common data processing patterns and how to choose the best data store for different parts of the data pipeline.
Accelerating a Path to Digital With a Cloud Data StrategyMongoDB
The document describes a conference on accelerating a path to digital transformation with a cloud data strategy. It provides an agenda for the conference including speakers on executing a cloud data strategy, customer stories from De Persgroep and Toyota Motor Europe, and a session on landing in the cloud with MongoDB Atlas. The document also provides background on the speakers and their companies.
MongoDB Partner Program Update - November 2013MongoDB
The document provides details about an upcoming webinar for the MongoDB Partner Program quarterly update in November 2013. It includes information about webinar logistics such as Q&A, recordings, audio connections. It then discusses the webinar presenters and provides a brief history and updates on the MongoDB Partner Program including growth in partners, new benefits for partners, and education resources. It concludes with the program roadmap and next steps for partners.
A Common Problem:
- My Reports run slow
- Reports take 3 hours to run
- We don’t have enough time to run our reports
- It takes 5 minutes to view the first page!
As the report processing time increases, so the frustration level.
MongoDB Ops Manager is the easiest way to manage/monitor/operationalize your MongoDB footprint across your enterprise. Ops Manager automates key operations such as deployments, scaling, upgrades, and backups, all with the click of a button and integration with your favorite tools. It also provide the ability to monitor and alert on dozens of platform specific metrics. In this webinar, we'll cover the components of Ops Manager, as well as how it integrates and accelerates your use of MongoDB.
MongoDB SoCal 2020: Migrate Anything* to MongoDB AtlasMongoDB
This presentation discusses migrating data from other data stores to MongoDB Atlas. It begins by explaining why MongoDB and Atlas are good choices for data management. Several preparation steps are covered, including sizing the target Atlas cluster, increasing the source oplog, and testing connectivity. Live migration, mongomirror, and dump/restore options are presented for migrating between replicasets or sharded clusters. Post-migration steps like monitoring and backups are also discussed. Finally, migrating from other data stores like AWS DocumentDB, Azure CosmosDB, DynamoDB, and relational databases are briefly covered.
MongoDB SoCal 2020: Go on a Data Safari with MongoDB Charts!MongoDB
These days, everyone is expected to be a data analyst. But with so much data available, how can you make sense of it and be sure you're making the best decisions? One great approach is to use data visualizations. In this session, we take a complex dataset and show how the breadth of capabilities in MongoDB Charts can help you turn bits and bytes into insights.
MongoDB SoCal 2020: Using MongoDB Services in Kubernetes: Any Platform, Devel...MongoDB
MongoDB Kubernetes operator and MongoDB Open Service Broker are ready for production operations. Learn about how MongoDB can be used with the most popular container orchestration platform, Kubernetes, and bring self-service, persistent storage to your containerized applications. A demo will show you how easy it is to enable MongoDB clusters as an External Service using the Open Service Broker API for MongoDB
MongoDB SoCal 2020: A Complete Methodology of Data Modeling for MongoDBMongoDB
Are you new to schema design for MongoDB, or are you looking for a more complete or agile process than what you are following currently? In this talk, we will guide you through the phases of a flexible methodology that you can apply to projects ranging from small to large with very demanding requirements.
MongoDB SoCal 2020: From Pharmacist to Analyst: Leveraging MongoDB for Real-T...MongoDB
Humana, like many companies, is tackling the challenge of creating real-time insights from data that is diverse and rapidly changing. This is our journey of how we used MongoDB to combined traditional batch approaches with streaming technologies to provide continues alerting capabilities from real-time data streams.
MongoDB SoCal 2020: Best Practices for Working with IoT and Time-series DataMongoDB
Time series data is increasingly at the heart of modern applications - think IoT, stock trading, clickstreams, social media, and more. With the move from batch to real time systems, the efficient capture and analysis of time series data can enable organizations to better detect and respond to events ahead of their competitors or to improve operational efficiency to reduce cost and risk. Working with time series data is often different from regular application data, and there are best practices you should observe.
This talk covers:
Common components of an IoT solution
The challenges involved with managing time-series data in IoT applications
Different schema designs, and how these affect memory and disk utilization – two critical factors in application performance.
How to query, analyze and present IoT time-series data using MongoDB Compass and MongoDB Charts
At the end of the session, you will have a better understanding of key best practices in managing IoT time-series data with MongoDB.
Join this talk and test session with a MongoDB Developer Advocate where you'll go over the setup, configuration, and deployment of an Atlas environment. Create a service that you can take back in a production-ready state and prepare to unleash your inner genius.
MongoDB .local San Francisco 2020: Powering the new age data demands [Infosys]MongoDB
Our clients have unique use cases and data patterns that mandate the choice of a particular strategy. To implement these strategies, it is mandatory that we unlearn a lot of relational concepts while designing and rapidly developing efficient applications on NoSQL. In this session, we will talk about some of our client use cases, the strategies we have adopted, and the features of MongoDB that assisted in implementing these strategies.
MongoDB .local San Francisco 2020: Using Client Side Encryption in MongoDB 4.2MongoDB
Encryption is not a new concept to MongoDB. Encryption may occur in-transit (with TLS) and at-rest (with the encrypted storage engine). But MongoDB 4.2 introduces support for Client Side Encryption, ensuring the most sensitive data is encrypted before ever leaving the client application. Even full access to your MongoDB servers is not enough to decrypt this data. And better yet, Client Side Encryption can be enabled at the "flick of a switch".
This session covers using Client Side Encryption in your applications. This includes the necessary setup, how to encrypt data without sacrificing queryability, and what trade-offs to expect.
MongoDB .local San Francisco 2020: Using MongoDB Services in Kubernetes: any ...MongoDB
MongoDB Kubernetes operator is ready for prime-time. Learn about how MongoDB can be used with most popular orchestration platform, Kubernetes, and bring self-service, persistent storage to your containerized applications.
MongoDB .local San Francisco 2020: Go on a Data Safari with MongoDB Charts!MongoDB
These days, everyone is expected to be a data analyst. But with so much data available, how can you make sense of it and be sure you're making the best decisions? One great approach is to use data visualizations. In this session, we take a complex dataset and show how the breadth of capabilities in MongoDB Charts can help you turn bits and bytes into insights.
MongoDB .local San Francisco 2020: From SQL to NoSQL -- Changing Your MindsetMongoDB
When you need to model data, is your first instinct to start breaking it down into rows and columns? Mine used to be too. When you want to develop apps in a modern, agile way, NoSQL databases can be the best option. Come to this talk to learn how to take advantage of all that NoSQL databases have to offer and discover the benefits of changing your mindset from the legacy, tabular way of modeling data. We’ll compare and contrast the terms and concepts in SQL databases and MongoDB, explain the benefits of using MongoDB compared to SQL databases, and walk through data modeling basics so you feel confident as you begin using MongoDB.
MongoDB .local San Francisco 2020: MongoDB Atlas JumpstartMongoDB
Join this talk and test session with a MongoDB Developer Advocate where you'll go over the setup, configuration, and deployment of an Atlas environment. Create a service that you can take back in a production-ready state and prepare to unleash your inner genius.
MongoDB .local San Francisco 2020: Tips and Tricks++ for Querying and Indexin...MongoDB
The document discusses guidelines for ordering fields in compound indexes to optimize query performance. It recommends the E-S-R approach: placing equality fields first, followed by sort fields, and range fields last. This allows indexes to leverage equality matches, provide non-blocking sorts, and minimize scanning. Examples show how indexes ordered by these guidelines can support queries more efficiently by narrowing the search bounds.
MongoDB .local San Francisco 2020: Aggregation Pipeline Power++MongoDB
Aggregation pipeline has been able to power your analysis of data since version 2.2. In 4.2 we added more power and now you can use it for more powerful queries, updates, and outputting your data to existing collections. Come hear how you can do everything with the pipeline, including single-view, ETL, data roll-ups and materialized views.
MongoDB .local San Francisco 2020: A Complete Methodology of Data Modeling fo...MongoDB
The document describes a methodology for data modeling with MongoDB. It begins by recognizing the differences between document and tabular databases, then outlines a three step methodology: 1) describe the workload by listing queries, 2) identify and model relationships between entities, and 3) apply relevant patterns when modeling for MongoDB. The document uses examples around modeling a coffee shop franchise to illustrate modeling approaches and techniques.
MongoDB .local San Francisco 2020: MongoDB Atlas Data Lake Technical Deep DiveMongoDB
MongoDB Atlas Data Lake is a new service offered by MongoDB Atlas. Many organizations store long term, archival data in cost-effective storage like S3, GCP, and Azure Blobs. However, many of them do not have robust systems or tools to effectively utilize large amounts of data to inform decision making. MongoDB Atlas Data Lake is a service allowing organizations to analyze their long-term data to discover a wealth of information about their business.
This session will take a deep dive into the features that are currently available in MongoDB Atlas Data Lake and how they are implemented. In addition, we'll discuss future plans and opportunities and offer ample Q&A time with the engineers on the project.
MongoDB .local San Francisco 2020: Developing Alexa Skills with MongoDB & GolangMongoDB
Virtual assistants are becoming the new norm when it comes to daily life, with Amazon’s Alexa being the leader in the space. As a developer, not only do you need to make web and mobile compliant applications, but you need to be able to support virtual assistants like Alexa. However, the process isn’t quite the same between the platforms.
How do you handle requests? Where do you store your data and work with it to create meaningful responses with little delay? How much of your code needs to change between platforms?
In this session we’ll see how to design and develop applications known as Skills for Amazon Alexa powered devices using the Go programming language and MongoDB.
MongoDB .local Paris 2020: Realm : l'ingrédient secret pour de meilleures app...MongoDB
aux Core Data, appréciée par des centaines de milliers de développeurs. Apprenez ce qui rend Realm spécial et comment il peut être utilisé pour créer de meilleures applications plus rapidement.
MongoDB .local Paris 2020: Upply @MongoDB : Upply : Quand le Machine Learning...MongoDB
Il n’a jamais été aussi facile de commander en ligne et de se faire livrer en moins de 48h très souvent gratuitement. Cette simplicité d’usage cache un marché complexe de plus de 8000 milliards de $.
La data est bien connu du monde de la Supply Chain (itinéraires, informations sur les marchandises, douanes,…), mais la valeur de ces données opérationnelles reste peu exploitée. En alliant expertise métier et Data Science, Upply redéfinit les fondamentaux de la Supply Chain en proposant à chacun des acteurs de surmonter la volatilité et l’inefficacité du marché.
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdfAbi john
Analyze the growth of meme coins from mere online jokes to potential assets in the digital economy. Explore the community, culture, and utility as they elevate themselves to a new era in cryptocurrency.
Spark is a powerhouse for large datasets, but when it comes to smaller data workloads, its overhead can sometimes slow things down. What if you could achieve high performance and efficiency without the need for Spark?
At S&P Global Commodity Insights, having a complete view of global energy and commodities markets enables customers to make data-driven decisions with confidence and create long-term, sustainable value. 🌍
Explore delta-rs + CDC and how these open-source innovations power lightweight, high-performance data applications beyond Spark! 🚀
Special Meetup Edition - TDX Bengaluru Meetup #52.pptxshyamraj55
We’re bringing the TDX energy to our community with 2 power-packed sessions:
🛠️ Workshop: MuleSoft for Agentforce
Explore the new version of our hands-on workshop featuring the latest Topic Center and API Catalog updates.
📄 Talk: Power Up Document Processing
Dive into smart automation with MuleSoft IDP, NLP, and Einstein AI for intelligent document workflows.
This is the keynote of the Into the Box conference, highlighting the release of the BoxLang JVM language, its key enhancements, and its vision for the future.
Procurement Insights Cost To Value Guide.pptxJon Hansen
Procurement Insights integrated Historic Procurement Industry Archives, serves as a powerful complement — not a competitor — to other procurement industry firms. It fills critical gaps in depth, agility, and contextual insight that most traditional analyst and association models overlook.
Learn more about this value- driven proprietary service offering here.
TrsLabs - Fintech Product & Business ConsultingTrs Labs
Hybrid Growth Mandate Model with TrsLabs
Strategic Investments, Inorganic Growth, Business Model Pivoting are critical activities that business don't do/change everyday. In cases like this, it may benefit your business to choose a temporary external consultant.
An unbiased plan driven by clearcut deliverables, market dynamics and without the influence of your internal office equations empower business leaders to make right choices.
Getting things done within a budget within a timeframe is key to Growing Business - No matter whether you are a start-up or a big company
Talk to us & Unlock the competitive advantage
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep DiveScyllaDB
Want to learn practical tips for designing systems that can scale efficiently without compromising speed?
Join us for a workshop where we’ll address these challenges head-on and explore how to architect low-latency systems using Rust. During this free interactive workshop oriented for developers, engineers, and architects, we’ll cover how Rust’s unique language features and the Tokio async runtime enable high-performance application development.
As you explore key principles of designing low-latency systems with Rust, you will learn how to:
- Create and compile a real-world app with Rust
- Connect the application to ScyllaDB (NoSQL data store)
- Negotiate tradeoffs related to data modeling and querying
- Manage and monitor the database for consistently low latencies
Quantum Computing Quick Research Guide by Arthur MorganArthur Morgan
This is a Quick Research Guide (QRG).
QRGs include the following:
- A brief, high-level overview of the QRG topic.
- A milestone timeline for the QRG topic.
- Links to various free online resource materials to provide a deeper dive into the QRG topic.
- Conclusion and a recommendation for at least two books available in the SJPL system on the QRG topic.
QRGs planned for the series:
- Artificial Intelligence QRG
- Quantum Computing QRG
- Big Data Analytics QRG
- Spacecraft Guidance, Navigation & Control QRG (coming 2026)
- UK Home Computing & The Birth of ARM QRG (coming 2027)
Any questions or comments?
- Please contact Arthur Morgan at [email protected].
100% human made.
Book industry standards are evolving rapidly. In the first part of this session, we’ll share an overview of key developments from 2024 and the early months of 2025. Then, BookNet’s resident standards expert, Tom Richardson, and CEO, Lauren Stewart, have a forward-looking conversation about what’s next.
Link to recording, presentation slides, and accompanying resource: https://bnctechforum.ca/sessions/standardsgoals-for-2025-standards-certification-roundup/
Presented by BookNet Canada on May 6, 2025 with support from the Department of Canadian Heritage.
#StandardsGoals for 2025: Standards & certification roundup - Tech Forum 2025BookNet Canada
Book industry standards are evolving rapidly. In the first part of this session, we’ll share an overview of key developments from 2024 and the early months of 2025. Then, BookNet’s resident standards expert, Tom Richardson, and CEO, Lauren Stewart, have a forward-looking conversation about what’s next.
Link to recording, transcript, and accompanying resource: https://bnctechforum.ca/sessions/standardsgoals-for-2025-standards-certification-roundup/
Presented by BookNet Canada on May 6, 2025 with support from the Department of Canadian Heritage.
Big Data Analytics Quick Research Guide by Arthur MorganArthur Morgan
This is a Quick Research Guide (QRG).
QRGs include the following:
- A brief, high-level overview of the QRG topic.
- A milestone timeline for the QRG topic.
- Links to various free online resource materials to provide a deeper dive into the QRG topic.
- Conclusion and a recommendation for at least two books available in the SJPL system on the QRG topic.
QRGs planned for the series:
- Artificial Intelligence QRG
- Quantum Computing QRG
- Big Data Analytics QRG
- Spacecraft Guidance, Navigation & Control QRG (coming 2026)
- UK Home Computing & The Birth of ARM QRG (coming 2027)
Any questions or comments?
- Please contact Arthur Morgan at [email protected].
100% human made.
What is Model Context Protocol(MCP) - The new technology for communication bw...Vishnu Singh Chundawat
The MCP (Model Context Protocol) is a framework designed to manage context and interaction within complex systems. This SlideShare presentation will provide a detailed overview of the MCP Model, its applications, and how it plays a crucial role in improving communication and decision-making in distributed systems. We will explore the key concepts behind the protocol, including the importance of context, data management, and how this model enhances system adaptability and responsiveness. Ideal for software developers, system architects, and IT professionals, this presentation will offer valuable insights into how the MCP Model can streamline workflows, improve efficiency, and create more intuitive systems for a wide range of use cases.
HCL Nomad Web – Best Practices and Managing Multiuser Environmentspanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-nomad-web-best-practices-and-managing-multiuser-environments/
HCL Nomad Web is heralded as the next generation of the HCL Notes client, offering numerous advantages such as eliminating the need for packaging, distribution, and installation. Nomad Web client upgrades will be installed “automatically” in the background. This significantly reduces the administrative footprint compared to traditional HCL Notes clients. However, troubleshooting issues in Nomad Web present unique challenges compared to the Notes client.
Join Christoph and Marc as they demonstrate how to simplify the troubleshooting process in HCL Nomad Web, ensuring a smoother and more efficient user experience.
In this webinar, we will explore effective strategies for diagnosing and resolving common problems in HCL Nomad Web, including
- Accessing the console
- Locating and interpreting log files
- Accessing the data folder within the browser’s cache (using OPFS)
- Understand the difference between single- and multi-user scenarios
- Utilizing Client Clocking
Generative Artificial Intelligence (GenAI) in BusinessDr. Tathagat Varma
My talk for the Indian School of Business (ISB) Emerging Leaders Program Cohort 9. In this talk, I discussed key issues around adoption of GenAI in business - benefits, opportunities and limitations. I also discussed how my research on Theory of Cognitive Chasms helps address some of these issues
Technology Trends in 2025: AI and Big Data AnalyticsInData Labs
At InData Labs, we have been keeping an ear to the ground, looking out for AI-enabled digital transformation trends coming our way in 2025. Our report will provide a look into the technology landscape of the future, including:
-Artificial Intelligence Market Overview
-Strategies for AI Adoption in 2025
-Anticipated drivers of AI adoption and transformative technologies
-Benefits of AI and Big data for your business
-Tips on how to prepare your business for innovation
-AI and data privacy: Strategies for securing data privacy in AI models, etc.
Download your free copy nowand implement the key findings to improve your business.
2. Themes
Broader use case portfolio. Pluggable storage engine strategy enables us to
rapidly cover more use cases with a single database.
Mission-critical apps. MongoDB delivers major advances in the critical areas
of governance, high availability, and disaster recovery.
New tools for new users. Now MongoDB is an integral part of the tooling and
workflows of Data Analysts, DBAs, and Operations teams.
4. Storage Engine Architecture in 3.2
Content
Repo
IoT Sensor
Backend
Ad Service
Customer
Analytics
Archive
MongoDB Query Language (MQL) + Native Drivers
MongoDB Document Data Model
Wired Tiger MMAP
Supported in MongoDB 3.2
Management
Security
In-memory
(beta)
Encrypted 3rd party
5. WiredTiger is the New Default
WiredTiger – widely deployed with 3.0 – is
now the default storage engine for
MongoDB.
• Best general purpose storage engine
• 7-10x better write throughput
• Up to 80% compression
6. Encrypted Storage Engine
Encrypted storage engine for end-to-end
encryption of sensitive data in regulated
industries
• Reduces the management and performance
overhead of external encryption mechanisms
• AES-256 Encryption, FIPS 140-2 option available
• Key management: Local key management via
keyfile or integration with 3rd party key
management appliance via KMIP
• Offered as an option for WiredTiger storage engine
7. In-Memory Storage Engine (Beta)
Handle ultra-high throughput with low
latency and high availability
• Delivers the extreme throughput and predictable
latency required by the most demanding apps in
Adtech, finance, and more.
• Achieve data durability with replica set members
running disk-backed storage engine
• Available for beta testing now and is expected for
GA in 2016
10. Data Governance with Document Validation
Implement data governance without
sacrificing agility that comes from dynamic
schema
• Enforce data quality across multiple teams and
applications
• Use familiar MongoDB expressions to control
document structure
• Validation is optional and can be as simple as a
single field, all the way to every field, including
existence, data types, and regular expressions
11. Document Validation Example
The example on the left adds a rule to the
contacts collection that validates:
• The year of birth is no later than 1994
• The document contains a phone number and / or
an email address
• When present, the phone number and email
addresses are strings
12. Enhancements for your mission-critical apps
More improvements in 3.2 that optimize the
database for your mission-critical
applications
• Meet stringent SLAs with fast-failover algorithm
– Under 2 seconds to detect and recover from
replica set primary failure
• Simplified management of sharded clusters
allow you to easily scale to many data centers
– Config servers are now deployed as replica
sets; up to 50 members
14. For Business Analysts & Data Scientists
MongoDB 3.2 allows business analysts and
data scientists to support the business with
new insights from untapped data sources
• MongoDB Connector for BI
• Dynamic Lookup
• New Aggregation Operators & Improved Text Search
15. MongoDB Connector for BI
Visualize and explore multi-dimensional
documents using SQL-based BI tools. The
connector does the following:
• Provides the BI tool with the schema of the
MongoDB collection to be visualized
• Translates SQL statements issued by the BI tool
into equivalent MongoDB queries that are sent to
MongoDB for processing
• Converts the results into the tabular format
expected by the BI tool, which can then visualize
the data based on user requirements
16. Dynamic Lookup
Combine data from multiple collections with
left outer joins for richer analytics & more
flexibility in data modeling
• Blend data from multiple sources for analysis
• Higher performance analytics with less application-
side code and less effort from your developers
• Executed via the new $lookup operator, a stage in
the MongoDB Aggregation Framework pipeline
17. Improved In-Database Analytics & Search
New Aggregation operators extend options for
performing analytics and ensure that answers
are delivered quickly and simply with lower
developer complexity
• Array operators: $slice, $arrayElemAt, $concatArrays,
$filter, $min, $max, $avg, $sum, and more
• New mathematical operators: $stdDevSamp,
$stdDevPop, $sqrt, $abs, $trunc, $ceil, $floor, $log,
$pow, $exp, and more
• Case sensitive text search and support for additional
languages such as Arabic, Farsi, Chinese, and more
18. For Operations Teams
MongoDB 3.2 simplifies and enhances
MongoDB’s management platforms. Ops
teams can be 10-20x more productive using
Ops and Cloud Manager to run MongoDB.
• Start from a global view of infrastructure:
Integrations with Application Performance
Monitoring platforms
• Drill down: Visual query performance diagnostics,
index recommendations
• Then, deploy: Automated index builds
• Refine: Partial indexes improve resource
utilization
19. Integrations with APM Platforms
Easily incorporate MongoDB performance
metrics into your existing APM dashboards
for global oversight of your entire IT stack
• MongoDB drivers enhanced with new API that
exposed query performance metrics to APM tools
• In addition, Ops and Cloud Manager can
complement this functionality with rich database
monitoring.
20. Query Perf. Visualizations & Optimization
Fast and simple query optimization with the
new Visual Query Profiler
• Query and write latency are consolidated and
displayed visually; your ops teams can easily
identify slower queries and latency spikes
• Visual query profiler analyzes the data it displays
and provides recommendations for new indexes
that can be created to improve query performance
• Ops Manager and Cloud Manager can automate
the rollout of new indexes, reducing risk and your
team’s operational overhead
21. Refine with Partial Indexes
Balance delivering good query performance
while consuming fewer system resources
• Specify a filtering expression during index creation
to instruct MongoDB to only include documents
that meet your desired conditions
• The example to the left creates an index that only
indexes the transaction documents with a modified
date since Jan 1, 2014
db.transactions.createIndex(
{modifiedDate: -1},
{partialFilterExpression:
{modifiedDate:
{$gt: Date("2014-01-01T00:00:00.000Z")}
}
}
)
22. For Database Administrators
MongoDB 3.2 helps users in your
organization understand the data in your
database
• MongoDB Compass
– For DBAs responsible for maintaining the
database in production
– No knowledge of the MongoDB query
language required
23. MongoDB Compass
For fast schema discovery and visual
construction of ad-hoc queries
• Visualize schema
– Frequency of fields
– Frequency of types
– Determine validator rules
• View Documents
• Graphically build queries
• Authenticated access
24. MongoDB
Enterprise Advanced
The best way to run MongoDB in your
data center.
Automated.
Supported.
Secured.
What’s included?
Enterprise-Grade Support
Encrypted & In-Memory Storage Engines
MongoDB Compass
BI Connector
Ops Manager or Cloud Manager Premium
Advanced Security
Commercial License
Platform Certification
On-Demand Training
Editor's Notes
#3: How many of you have already upgraded to MongoDB version 3.2?
There are 3 basic themes for the 3.2 release each of which we will discuss in detail:
Broadening our use case portfolio
Better support for Mission Critical-apps
And, New tools for new users
#5: MongoDB 3.0 introduced a new flexible storage engine architecture. This makes it fast and easy to build new pluggable storage engines to extend the database with new capabilities and address specific workload requirements.
MongoDB 3.2 adds 2 new options to the mix:
An Encrypted storage engine to protect highly sensitive data, without the performance or management overhead of separate filesystem encryption.
And an In-memory storage engine that delivers extreme performance and predictable latency.
#6: The WiredTiger storage engine was first made available as an option with MongoDB 3.0. How many of you have moved to the WiredTiger storage engine?
With the 3.2 version of MongoDB, WiredTiger as its default storage engine.
Fresh download would start up with WiredTiger as the storage engine
When compared to the original MMAP storage engine used in earlier MongoDB releases, write performance has improved by 7 to 10 times, and storage overhead has reduced by up to 80%. These improvements were made possible due to WiredTiger’s more granular concurrency control and native compression.
#7: With the introduction of the Encrypted storage engine, MongoDB now provides end-to-end encryption.
Because the storage engine natively encrypts the database files on disk, the management and performance overhead of other external encryption mechanisms is reduced.
We use 256-bit encryption, and data is encrypted using an algorithm that takes a random encryption key as input and generates ciphertext that can only be read if decrypted with the decryption key. This encryption/decryption process is completely transparent to the application, and the storage engine encrypts each database with a separate key.
MongoDB supports two key management options:
Local key management via a key file OR
Integration with the 3rd party key management appliance using the KMIP protocol.
#8: Most people understand the advantages of in-memory computing. Data can be accessed in RAM nearly 100,000 times faster than retrieving it from disk, and writes do not need to be persisted to disk. Because of this fast access and the fact that the data never needs to be persisted to disk, the in-memory storage engine delivers the highest possible performance for the most demanding applications.
Data durability and persistence can be achieved by using secondary replica set members with disk-based storage.
The in-memory storage engine is available for beta testing now and is expect to be General Availability this year (2016).
#9: In this ecommerce example, we are using a mix of storage engines. User data is managed by the In-Memory engine to provide the throughput and low latency needed for great customer experience, while the product catalog’s data gets provisioned to another MongoDB replica set configured with the disk-based WiredTiger storage engine.
You’ll also notice that in the user data replica set example that although the primary is using the in-memory storage engine, the secondaries are using the WiredTiger storage engine with disk-based storage. So even though no data gets persisted to disk on the primary, the secondaries provide additional data protection by persisting the data to disk.
Our flexible storage architecture means there is no need to use different in-memory and disk-based databases to support the e-commerce application. It’s the same Mongo database for both replica sets, it’s just a different storage engine that happens to be the best suited for the use case.
#11: We all know that one of the major benefits of MongoDB is it’s dynamic and flexible schema, but it is also important to be able to implement controls to maintain data quality. To address this, MongoDB no offers document validation within the database so that data validation doesn’t need to take place within the application.
Validation checks can be done on document structure, data types, data ranges, and the presence of mandatory fields.
Validations can be set to reject documents that violate the validation rule OR warn and log the violation but allow the invalid document to be inserted or updated
#12: Just to show you an example of how to add a validation, here we are adding a rule to the “CONTACTS” collection. For any newly inserted or updated documents, this rule will check that the year of birth is less than or equal to 1994, that it contains a phone number AND/OR an email address, and that when the phone number or email address fields are present, that they are a string data type.
#13: MongoDB 3.2 also introduced an enhanced replication protocol available that delivers faster recovery in the event of a primary failure. The enhanced protocol reduces the length of time for failover by optimizing the algorithm to detect replica set primary failures and elect a new primary. Failover time is dependent on several factors including network latency.
For those of you with sharded clusters, another enhancement with MongoDB 3.2 is that config servers can now be deployed as replica sets. Prior to 3.2, the config servers were implemented as 3 AND ONLY 3 special-purpose servers with their own write protocols, coordinators, and consistency checking.
#15: We’ve added new capabilities targeted for use by business analysts and data scientists
#16: SQL-based BI tools expect to connect to a data source with a fixed schema presenting tabular data. This can be a challenge when working with MongoDB’s dynamic schema and rich documents. To allow SQL-based BI tools to query MongoDB as a data source, we have developed the BI Connector. Now, business analysts can use BI visualization tools like Tableau, Qlik, Business Objects, and Cognos.
The BI Connector performs 3 functions:
First, it provides the BI tool with the schema of the MongoDB collection. The schema output generated by the BI Connector can be reviewed to ensure that data types, sub-documents and arrays are represented correctly.
Second, it translates SQL statements issued by the BI tool into the equivalent MongoDB queries that are then sent to MongoDB for processing.
And lastly, it converts the returned results into the tabular format expected by the BI tool, which can then visualize the data based on the user requirements.
#17: We’ve also added the ability to combine data from multiple collections by implementing left outer joins via the $lookup operator. The $lookup operator can be included as a stage in the MongoDB aggregation framework and gives you more flexibility in data modeling. It allows richer analytics to be run in the database with higher performance and less application-side code.
#18: MongoDB 3.2 also expands the options for running analytics on the operational database.
We already know that MongoDB has a very flexible schema, and we can store arrays as well as simple values. Being able to manipulate and filter these arrays during aggregation is really important. New operators have been added to allow for more flexibility when dealing with arrays.
We’ve also added new mathematical operators such as truncate, ceiling, floor, absolute, and standard deviation. These operators allow you to move code out of the client tier directly into the database resulting in higher performance and less developer complexity.
And for text searches, we have increased the set of use cases by adding support for case-sensitive searches, as well as, additional languages.
#19: We’ve made several enhancements that are applicable to Operations teams for monitoring and managing MongoDB
#20: Many operations teams use Application Performance Management platforms like New Relic and AppDynamics to view their entire IT infrastructure from a single “pane of glass”. We now have a new API that exposes query performance metrics to these APM tools.
Also, MongoDB Cloud Manager now provides packaged integration with the New Relic platform. Nothing additional needs to be installed, you can just copy and paste a couple of keys to display the MongoDB metrics, events, and visualizations directly in New Relic.
#21: Some of you may already be intimately familiar with the database profiler. The profiler collects fined-grained information that can be used to analyze query performance. Parsing through the profiler output, though, isn’t always the easiest thing to do. Ops Manager and Cloud Manager Premium now include a Visual Query Profiler that gives DBAs and operations teams a quick and easy way to analyze specific queries.
It visually displays how query and write latency varies over time and helps make it easy to identify slow queries and latency spikes.
The Visual Query Profiler also analyzes the data it collects and can make recommendations for new indexes to improve query performance.
#22: We all know that Secondary Indexes are one of the ways that MongoDB is different from other NoSQL databases. They help provide efficient access to data, and increase read performance.
BUT.....indexes do have a cost: 1) Databases writes will be slower when they need to update the index 2) Memory and storage are needed to store the index.
Partial indexes are meant to balance the delivery of good query performance while using fewer systems resources.
By specifying a filtering expression when creating an index – like having a date greater than January 1, 2014, only documents that meet that criteria will be included in the index.
#23: And last but not least for you Database Administrators, we’ve also create a new schema visualization and ad-hoc query tool.
#24: In the past, it has been difficult to explore and understand the underlying data and its structure in MongoDB because of our dynamic schema. It meant connecting to the MongoDB shell and writing queries to determine the document structure, field names, and data types. Now, we have a new graphical tool called MongoDB Compass. Compass allows users to understand the structure of the data in the database and perform ad-hoc queries without having to understand MongoDB’s query language.
The tool works by sampling a subset of documents from a collection and it displays information about the schema such as frequency of fields and frequency of types. By using a sampling algorithm, it minimizes overhead and can present results to the user almost instantly.
To query a collection, document elements can be selected from the user interface and the query can then be run with the push of a button. Results from the query can be viewed both graphically and as a set of JSON documents.
And now, one last slide, and then I’ll help you “find your way” with an actual MongoDB Compass demonstration.
#25: Many of you are probably using the open source community version of MongoDB, and that’s great, but MongoDB also offer an Enterprise version of our software that includes 7x24 support, a commercial license, a number of security features and capabilities like an encrypted storage engine, as well as operational management tools like Ops Manager or Cloud Manager.
7x24x365 Support with a 1 hr SLA
Encrypted Storage Engine for end-to-end database encryption
In memory storage engine for your ultra throughput, most demanding apps: in memory computing without sacrificing data durability
MongoDB Compass – Schema and data visualization; understand the data stored in your database with no knowledge of the MongoDB query language. Ad hoc queries with a few clicks of your mouse
BI Connector – Visualize and analyze the multi-structured data stored in MongoDB using SQL-based BI tools such as Tableau, Qlikview, Spotfire and more
MongoDB Ops Manager – full management platform to de-risk MongoDB in production
Monitor the health of your system
Automate deployment, configuration, maintenance, upgrades and scaling
Back up and restore to any point in time (standard network mountable filesystems supported)
Visual Query profiler to identify slow-running queries
Index suggestions and automated index rollouts
APM integration with enhanced drivers
Runs behind your firewall.
Enterprise-grade, follow the sun support with a 1-hour SLA
Not just break/fix support
Direct access to industry best-practices
Advanced Security
LDAP and Kerberos to integrate with existing authentication and authorization infrastructure
Auditing of all database operations for compliance
Commercial license
To meet the needs of organizations that have policies against using open source, AGPL software
Platform Certification
Tested and certified for stability and performance on Windows, Red Hat/CentOS, Ubuntu, and Amazon Linux
On-Demand Training
Access to our online courses at your own pace to get team members up to speed