Preview of slides on options for running the Oracle BI Applications "in-memory" using TimesTen for Exalytics, delivered at the UKOUG Analytics Event in London, July 2013
What is Big Data Discovery, and how it complements traditional business anal...Mark Rittman
Data Discovery is an analysis technique that complements traditional business analytics, and enables users to combine, explore and analyse disparate datasets to spot opportunities and patterns that lie hidden within your data. Oracle Big Data discovery takes this idea and applies it to your unstructured and big data datasets, giving users a way to catalogue, join and then analyse all types of data across your organization.
In this session we'll look at Oracle Big Data Discovery and how it provides a "visual face" to your big data initatives, and how it complements and extends the work that you currently do using business analytics tools.
Unlock the value in your big data reservoir using oracle big data discovery a...Mark Rittman
The document discusses Oracle Big Data Discovery and how it can be used to analyze and gain insights from data stored in a Hadoop data reservoir. It provides an example scenario where Big Data Discovery is used to analyze website logs, tweets, and website posts and comments to understand popular content and influencers for a company. The data is ingested into the Big Data Discovery tool, which automatically enriches the data. Users can then explore the data, apply additional transformations, and visualize relationships to gain insights.
OBIEE12c and Embedded Essbase 12c - An Initial Look at Query Acceleration Use...Mark Rittman
OBIEE12c comes with an updated version of Essbase that focuses entirely in this release on the query acceleration use-case. This presentation looks at this new release and explains how the new BI Accelerator Wizard manages the creation of Essbase cubes to accelerate OBIEE query performance
Big Data for Oracle Devs - Towards Spark, Real-Time and Predictive AnalyticsMark Rittman
This is a session for Oracle DBAs and devs that looks at the cutting edge big data techs like Spark, Kafka etc, and through demos shows how Hadoop is now a a real-time platform for fast analytics, data integration and predictive modeling
How To Leverage OBIEE Within A Big Data ArchitectureKevin McGinley
If you've invested in OBIEE and want to start exploring the use of Big Data technology, this presentation talks about how and why you might want to use OBIEE as the common visualization layer across both.
GoldenGate and ODI - A Perfect Match for Real-Time Data WarehousingMichael Rainey
Oracle Data Integrator and Oracle GoldenGate excel as standalone products, but paired together they are the perfect match for real-time data warehousing. Following Oracle’s Next Generation Reference Data Warehouse Architecture, this discussion will provide best practices on how to configure, implement, and process data in real-time using ODI and GoldenGate. Attendees will see common real-time challenges solved, including parent-child relationships within micro-batch ETL.
Presented at RMOUG Training Days 2013 & KScope13.
Real-time Data Warehouse Upgrade – Success StoriesMichael Rainey
Providing a real-time BI solution for its global customers and operations department is a necessity for IFPI, the International Federation of the Phonographic Industry, whose primary objective is to safeguard the rights of record producers through various anti-piracy strategies.
For the data warehousing team at IFPI, using Oracle Streams and Oracle Warehouse Builder (OWB) for real-time data replication and integration was becoming a challenge. The solution was difficult to maintain and overall throughput was degrading as data volumes increased. The need for greater stability and performance led IFPI to implement Oracle GoldenGate and Oracle Data Integrator.
Co-presented with Nick Hurt at Rittman Mead BI Forum 2014 and KScope14.
Riga dev day 2016 adding a data reservoir and oracle bdd to extend your ora...Mark Rittman
This talk focus is on what a data reservoir is, how it related to the RDBMS DW, and how Big Data Discovery provides access to it to business and BI users
Practical Tips for Oracle Business Intelligence Applications 11g ImplementationsMichael Rainey
The document provides practical tips for Oracle Business Intelligence Applications 11g implementations. It discusses scripting installations and configurations, LDAP integration challenges, implementing high availability, different methods for data extracts, and simplifying disaster recovery. Specific tips include scripting all processes, configuring the ODI agent JVM and connection pools for performance, understanding external LDAP authentication in ODI, implementing active-active high availability for ODI agents, choosing the right data extract method based on latency and volume, and using DataGuard and CNAMEs to simplify failover for disaster recovery.
Using Oracle Big Data SQL 3.0 to add Hadoop & NoSQL to your Oracle Data Wareh...Mark Rittman
As presented at OGh SQL Celebration Day in June 2016, NL. Covers new features in Big Data SQL including storage indexes, storage handlers and ability to install + license on commodity hardware
OBIEE & Essbase Integration with Oracle BI Foundation 11.1.1.7 (ODTUG 2013)Mark Rittman
A presentation on the architectural and installation changes for Essbase with the new 11.1.1.7 release of Oracle BI Foundation, including integrated security, the role Essbase plays in this new architecture, Essbase cube spin-off, and Smartview.
KScope14 - Real-Time Data Warehouse Upgrade - Success StoriesMichael Rainey
Providing real-time data to its global customers is a necessity for IFPI (International Federation of the Phonographic Industry), a not-for-profit organization with a mission to safeguard the rights of record producers and promote the value of recorded music. Using Oracle Streams and Oracle Warehouse Builder (OWB) for real-time data replication and integration, meeting this goal was becoming a challenge. The solution was difficult to maintain and overall throughput was degrading as data volume increased. The need for greater stability and performance led IFPI to implement Oracle GoldenGate and Oracle Data Integrator. This session will describe the innovative approach taken to complete the migration from a Streams and OWB implementation to a more robust, maintainable, and performant GoldenGate and ODI integrated solution.
In search of database nirvana - The challenges of delivering Hybrid Transacti...Rohit Jain
Companies are looking for a single database engine that can address all their varied needs—from transactional to analytical workloads, against structured, semi-structured, and unstructured data, leveraging graph, document, text search, column, key value, wide column, and relational data stores; on a single platform without the latency of data transformation and replication. They are looking for the ultimate database nirvana.
The term hybrid transactional/analytical processing (HTAP), coined by Gartner, perhaps comes closest to describing this concept. 451 Research uses the terms convergence or converged data platform. The terms multi-model or unified are also used. But can such a nirvana be achieved? Some database vendors claim to have already achieved this nirvana. In this talk we will discuss the following challenges on the path to this nirvana, for you to assess how accurate these claims are:
· What is needed for a single query engine to support all workloads?
· What does it take for that single query engine to support multiple storage engines, each serving a different need?
· Can a single query engine support all data models?
· Can it provide enterprise-caliber capabilities?
Attendees looking to assess query and storage engines would benefit from understanding what the key considerations are when picking an engine to run their targeted workloads. Also, developers working on such engines can better understand capabilities they need to provide in order to run workloads that span the HTAP spectrum.
End to-end hadoop development using OBIEE, ODI, Oracle Big Data SQL and Oracl...Mark Rittman
This document discusses an end-to-end example of using Hadoop, OBIEE, ODI and Oracle Big Data Discovery to analyze big data from various sources. It describes ingesting website log data and Twitter data into a Hadoop cluster, processing and transforming the data using tools like Hive and Spark, and using the results for reporting in OBIEE and data discovery in Oracle Big Data Discovery. ODI is used to automate the data integration process.
A Walk Through the Kimball ETL Subsystems with Oracle Data Integration - Coll...Michael Rainey
Big Data integration is an excellent feature in the Oracle Data Integration product suite (Oracle Data Integrator, GoldenGate, & Enterprise Data Quality). But not all analytics require big data technologies, such as labor cost, revenue, or expense reporting. Ralph Kimball, an original architect of the dimensional model in data warehousing, spent much of his career working to build an enterprise data warehouse methodology that can meet these reporting needs. His book, "The Data Warehouse ETL Toolkit", is a guide for many ETL developers. This session will walk you through his ETL Subsystem categories; Extracting, Cleaning & Conforming, Delivering, and Managing, describing how the Oracle Data Integration products are perfectly suited for the Kimball approach.
Presented at Collaborate16 in Las Vegas.
A Walk Through the Kimball ETL Subsystems with Oracle Data IntegrationMichael Rainey
Big Data integration is an excellent feature in the Oracle Data Integration product suite (Oracle Data Integrator, GoldenGate, & Enterprise Data Quality). But not all analytics require big data technologies, such as labor cost, revenue, or expense reporting. Ralph Kimball, an original architect of the dimensional model in data warehousing, spent much of his career working to build an enterprise data warehouse methodology that can meet these reporting needs. His book, "The Data Warehouse ETL Toolkit", is a guide for many ETL developers. This session will walk you through his ETL Subsystem categories; Extracting, Cleaning & Conforming, Delivering, and Managing, describing how the Oracle Data Integration products are perfectly suited for the Kimball approach.
Presented at Oracle OpenWorld 2015 & BIWA Summit 2016.
Adding a Data Reservoir to your Oracle Data Warehouse for Customer 360-Degree...Mark Rittman
This document summarizes a presentation about adding a Hadoop-based data reservoir to an Oracle data warehouse. The presentation discusses using a data reservoir to store large amounts of raw customer data from various sources to enable 360-degree customer analysis. It describes loading and integrating the data reservoir with the data warehouse using Oracle tools and how organizations can use it for more personalized customer marketing through advanced analytics and machine learning.
PyData: The Next Generation | Data Day Texas 2015Cloudera, Inc.
This document discusses the past, present, and future of Python for big data analytics. It provides background on the rise of Python as a data analysis tool through projects like NumPy, pandas, and scikit-learn. However, as big data systems like Hadoop became popular, Python was not initially well-suited for problems at that scale. Recent projects like PySpark, Blaze, and Spartan aim to bring Python to big data, but challenges remain around data formats, distributed computing interfaces, and competing with Scala. The document calls for continued investment in high performance Python tools for big data to ensure its relevance in coming years.
Part 1 - Introduction to Hadoop and Big Data Technologies for Oracle BI & DW ...Mark Rittman
Delivered as a one-day seminar at the SIOUG and HROUG Oracle User Group Conferences, October 2014
In this presentation we cover some key Hadoop concepts including HDFS, MapReduce, Hive and NoSQL/HBase, with the focus on Oracle Big Data Appliance and Cloudera Distribution including Hadoop. We explain how data is stored on a Hadoop system and the high-level ways it is accessed and analysed, and outline Oracle’s products in this area including the Big Data Connectors, Oracle Big Data SQL, and Oracle Business Intelligence (OBI) and Oracle Data Integrator (ODI).
Gluent New World #02 - SQL-on-Hadoop : A bit of History, Current State-of-the...Mark Rittman
Hadoop and NoSQL platforms initially focused on Java developers and slow but massively-scalable MapReduce jobs as an alternative to high-end but limited-scale analytics RDBMS engines. Apache Hive opened-up Hadoop to non-programmers by adding a SQL query engine and relational-style metadata layered over raw HDFS storage, and since then open-source initiatives such as Hive Stinger, Cloudera Impala and Apache Drill along with proprietary solutions from closed-source vendors have extended SQL-on-Hadoop’s capabilities into areas such as low-latency ad-hoc queries, ACID-compliant transactions and schema-less data discovery – at massive scale and with compelling economics.
In this session we’ll focus on technical foundations around SQL-on-Hadoop, first reviewing the basic platform Apache Hive provides and then looking in more detail at how ad-hoc querying, ACID-compliant transactions and data discovery engines work along with more specialised underlying storage that each now work best with – and we’ll take a look to the future to see how SQL querying, data integration and analytics are likely to come together in the next five years to make Hadoop the default platform running mixed old-world/new-world analytics workloads.
Social Network Analysis using Oracle Big Data Spatial & Graph (incl. why I di...Mark Rittman
The document discusses using Hadoop and NoSQL technologies like Apache HBase to perform social network analysis on Twitter data related to a company's website and blog. It describes ingesting tweet and website log data into Hadoop HDFS and processing it with tools like Hive. Graph algorithms from Oracle Big Data Spatial & Graph were then used on the property graph stored in HBase to identify influential Twitter users and communities. This approach provided real-time insights at scale compared to using a traditional relational database.
Integrating Oracle Data Integrator with Oracle GoldenGate 12cEdelweiss Kammermann
The document discusses integrating Oracle Data Integrator (ODI) with Oracle GoldenGate (OGG) for real-time data integration. It describes how OGG captures change data from source systems and delivers it to ODI. Key steps include configuring OGG installations and JAgents, defining OGG data servers in ODI, applying journalizing to ODI models, and creating and starting ODI processes that integrate with the OGG capture and delivery processes. The integration provides benefits like low impact on sources, great performance for real-time integration, and support for heterogeneous databases.
Reaching scale limits on a Hadoop platform: issues and errors created by spee...DataWorks Summit
Santander UK’s Big Data journey began in 2014, using Hadoop to make the most of our data and generate value for customers. Within 9 months, we created a highly available real-time customer facing application for customer analytics. We currently have 500 different people doing their own analysis and projects with this data, spanning a total of 50 different use cases. This data, (consisting of over 40 million customer records with billions of transactions), provides our business new insights that were inaccessible before.
Our business moves quickly, with several products and 20 use cases currently in production. We currently have a customer data lake and a technical data lake. Having a platform with very different workloads has proven to be challenging.
Our success in generating value created such growth in terms of data, use cases, analysts and usage patterns that 3 years later we find issues with scalability in HDFS, Hive metastore and Hadoop operations and challenges with highly available architectures with Hbase, Flume and Kafka. Going forward we are exploring alternative architectures including a hybrid cloud model, and moving towards streaming.
Our goal with this session is to assist people in the early part of their journey by building a solid foundation. We hope that others can benefit from us sharing our experiences and lessons learned during our journey.
Speaker
Nicolette Bullivant, Head of Data Engineering at Santander UK Technology, Santander UK Technology
Oracle Exalytics - Tips and Experiences from the Field (Enkitec E4 Conference...Mark Rittman
Presentation by Rittman Mead's Mark Rittman and Stewart Bryson on our experiences 1-year on with Exalytics. Includes sections on aggregate caching and datamart loading into TT, use of Essbase as a TT alternative, and deployment patterns we see on client sites.
OBIEE, Endeca, Hadoop and ORE Development (on Exalytics) (ODTUG 2013)Mark Rittman
A presentation from ODTUG 2013 on tools other than OBIEE for Exalytics, focusing on analysis of non-traditional data via Endeca, "big data" via Hadoop and statistical analysis / predictive modeling through Oracle R Enterprise, and the benefits of running these tools on Oracle Exalytics
Real-time Data Warehouse Upgrade – Success StoriesMichael Rainey
Providing a real-time BI solution for its global customers and operations department is a necessity for IFPI, the International Federation of the Phonographic Industry, whose primary objective is to safeguard the rights of record producers through various anti-piracy strategies.
For the data warehousing team at IFPI, using Oracle Streams and Oracle Warehouse Builder (OWB) for real-time data replication and integration was becoming a challenge. The solution was difficult to maintain and overall throughput was degrading as data volumes increased. The need for greater stability and performance led IFPI to implement Oracle GoldenGate and Oracle Data Integrator.
Co-presented with Nick Hurt at Rittman Mead BI Forum 2014 and KScope14.
Riga dev day 2016 adding a data reservoir and oracle bdd to extend your ora...Mark Rittman
This talk focus is on what a data reservoir is, how it related to the RDBMS DW, and how Big Data Discovery provides access to it to business and BI users
Practical Tips for Oracle Business Intelligence Applications 11g ImplementationsMichael Rainey
The document provides practical tips for Oracle Business Intelligence Applications 11g implementations. It discusses scripting installations and configurations, LDAP integration challenges, implementing high availability, different methods for data extracts, and simplifying disaster recovery. Specific tips include scripting all processes, configuring the ODI agent JVM and connection pools for performance, understanding external LDAP authentication in ODI, implementing active-active high availability for ODI agents, choosing the right data extract method based on latency and volume, and using DataGuard and CNAMEs to simplify failover for disaster recovery.
Using Oracle Big Data SQL 3.0 to add Hadoop & NoSQL to your Oracle Data Wareh...Mark Rittman
As presented at OGh SQL Celebration Day in June 2016, NL. Covers new features in Big Data SQL including storage indexes, storage handlers and ability to install + license on commodity hardware
OBIEE & Essbase Integration with Oracle BI Foundation 11.1.1.7 (ODTUG 2013)Mark Rittman
A presentation on the architectural and installation changes for Essbase with the new 11.1.1.7 release of Oracle BI Foundation, including integrated security, the role Essbase plays in this new architecture, Essbase cube spin-off, and Smartview.
KScope14 - Real-Time Data Warehouse Upgrade - Success StoriesMichael Rainey
Providing real-time data to its global customers is a necessity for IFPI (International Federation of the Phonographic Industry), a not-for-profit organization with a mission to safeguard the rights of record producers and promote the value of recorded music. Using Oracle Streams and Oracle Warehouse Builder (OWB) for real-time data replication and integration, meeting this goal was becoming a challenge. The solution was difficult to maintain and overall throughput was degrading as data volume increased. The need for greater stability and performance led IFPI to implement Oracle GoldenGate and Oracle Data Integrator. This session will describe the innovative approach taken to complete the migration from a Streams and OWB implementation to a more robust, maintainable, and performant GoldenGate and ODI integrated solution.
In search of database nirvana - The challenges of delivering Hybrid Transacti...Rohit Jain
Companies are looking for a single database engine that can address all their varied needs—from transactional to analytical workloads, against structured, semi-structured, and unstructured data, leveraging graph, document, text search, column, key value, wide column, and relational data stores; on a single platform without the latency of data transformation and replication. They are looking for the ultimate database nirvana.
The term hybrid transactional/analytical processing (HTAP), coined by Gartner, perhaps comes closest to describing this concept. 451 Research uses the terms convergence or converged data platform. The terms multi-model or unified are also used. But can such a nirvana be achieved? Some database vendors claim to have already achieved this nirvana. In this talk we will discuss the following challenges on the path to this nirvana, for you to assess how accurate these claims are:
· What is needed for a single query engine to support all workloads?
· What does it take for that single query engine to support multiple storage engines, each serving a different need?
· Can a single query engine support all data models?
· Can it provide enterprise-caliber capabilities?
Attendees looking to assess query and storage engines would benefit from understanding what the key considerations are when picking an engine to run their targeted workloads. Also, developers working on such engines can better understand capabilities they need to provide in order to run workloads that span the HTAP spectrum.
End to-end hadoop development using OBIEE, ODI, Oracle Big Data SQL and Oracl...Mark Rittman
This document discusses an end-to-end example of using Hadoop, OBIEE, ODI and Oracle Big Data Discovery to analyze big data from various sources. It describes ingesting website log data and Twitter data into a Hadoop cluster, processing and transforming the data using tools like Hive and Spark, and using the results for reporting in OBIEE and data discovery in Oracle Big Data Discovery. ODI is used to automate the data integration process.
A Walk Through the Kimball ETL Subsystems with Oracle Data Integration - Coll...Michael Rainey
Big Data integration is an excellent feature in the Oracle Data Integration product suite (Oracle Data Integrator, GoldenGate, & Enterprise Data Quality). But not all analytics require big data technologies, such as labor cost, revenue, or expense reporting. Ralph Kimball, an original architect of the dimensional model in data warehousing, spent much of his career working to build an enterprise data warehouse methodology that can meet these reporting needs. His book, "The Data Warehouse ETL Toolkit", is a guide for many ETL developers. This session will walk you through his ETL Subsystem categories; Extracting, Cleaning & Conforming, Delivering, and Managing, describing how the Oracle Data Integration products are perfectly suited for the Kimball approach.
Presented at Collaborate16 in Las Vegas.
A Walk Through the Kimball ETL Subsystems with Oracle Data IntegrationMichael Rainey
Big Data integration is an excellent feature in the Oracle Data Integration product suite (Oracle Data Integrator, GoldenGate, & Enterprise Data Quality). But not all analytics require big data technologies, such as labor cost, revenue, or expense reporting. Ralph Kimball, an original architect of the dimensional model in data warehousing, spent much of his career working to build an enterprise data warehouse methodology that can meet these reporting needs. His book, "The Data Warehouse ETL Toolkit", is a guide for many ETL developers. This session will walk you through his ETL Subsystem categories; Extracting, Cleaning & Conforming, Delivering, and Managing, describing how the Oracle Data Integration products are perfectly suited for the Kimball approach.
Presented at Oracle OpenWorld 2015 & BIWA Summit 2016.
Adding a Data Reservoir to your Oracle Data Warehouse for Customer 360-Degree...Mark Rittman
This document summarizes a presentation about adding a Hadoop-based data reservoir to an Oracle data warehouse. The presentation discusses using a data reservoir to store large amounts of raw customer data from various sources to enable 360-degree customer analysis. It describes loading and integrating the data reservoir with the data warehouse using Oracle tools and how organizations can use it for more personalized customer marketing through advanced analytics and machine learning.
PyData: The Next Generation | Data Day Texas 2015Cloudera, Inc.
This document discusses the past, present, and future of Python for big data analytics. It provides background on the rise of Python as a data analysis tool through projects like NumPy, pandas, and scikit-learn. However, as big data systems like Hadoop became popular, Python was not initially well-suited for problems at that scale. Recent projects like PySpark, Blaze, and Spartan aim to bring Python to big data, but challenges remain around data formats, distributed computing interfaces, and competing with Scala. The document calls for continued investment in high performance Python tools for big data to ensure its relevance in coming years.
Part 1 - Introduction to Hadoop and Big Data Technologies for Oracle BI & DW ...Mark Rittman
Delivered as a one-day seminar at the SIOUG and HROUG Oracle User Group Conferences, October 2014
In this presentation we cover some key Hadoop concepts including HDFS, MapReduce, Hive and NoSQL/HBase, with the focus on Oracle Big Data Appliance and Cloudera Distribution including Hadoop. We explain how data is stored on a Hadoop system and the high-level ways it is accessed and analysed, and outline Oracle’s products in this area including the Big Data Connectors, Oracle Big Data SQL, and Oracle Business Intelligence (OBI) and Oracle Data Integrator (ODI).
Gluent New World #02 - SQL-on-Hadoop : A bit of History, Current State-of-the...Mark Rittman
Hadoop and NoSQL platforms initially focused on Java developers and slow but massively-scalable MapReduce jobs as an alternative to high-end but limited-scale analytics RDBMS engines. Apache Hive opened-up Hadoop to non-programmers by adding a SQL query engine and relational-style metadata layered over raw HDFS storage, and since then open-source initiatives such as Hive Stinger, Cloudera Impala and Apache Drill along with proprietary solutions from closed-source vendors have extended SQL-on-Hadoop’s capabilities into areas such as low-latency ad-hoc queries, ACID-compliant transactions and schema-less data discovery – at massive scale and with compelling economics.
In this session we’ll focus on technical foundations around SQL-on-Hadoop, first reviewing the basic platform Apache Hive provides and then looking in more detail at how ad-hoc querying, ACID-compliant transactions and data discovery engines work along with more specialised underlying storage that each now work best with – and we’ll take a look to the future to see how SQL querying, data integration and analytics are likely to come together in the next five years to make Hadoop the default platform running mixed old-world/new-world analytics workloads.
Social Network Analysis using Oracle Big Data Spatial & Graph (incl. why I di...Mark Rittman
The document discusses using Hadoop and NoSQL technologies like Apache HBase to perform social network analysis on Twitter data related to a company's website and blog. It describes ingesting tweet and website log data into Hadoop HDFS and processing it with tools like Hive. Graph algorithms from Oracle Big Data Spatial & Graph were then used on the property graph stored in HBase to identify influential Twitter users and communities. This approach provided real-time insights at scale compared to using a traditional relational database.
Integrating Oracle Data Integrator with Oracle GoldenGate 12cEdelweiss Kammermann
The document discusses integrating Oracle Data Integrator (ODI) with Oracle GoldenGate (OGG) for real-time data integration. It describes how OGG captures change data from source systems and delivers it to ODI. Key steps include configuring OGG installations and JAgents, defining OGG data servers in ODI, applying journalizing to ODI models, and creating and starting ODI processes that integrate with the OGG capture and delivery processes. The integration provides benefits like low impact on sources, great performance for real-time integration, and support for heterogeneous databases.
Reaching scale limits on a Hadoop platform: issues and errors created by spee...DataWorks Summit
Santander UK’s Big Data journey began in 2014, using Hadoop to make the most of our data and generate value for customers. Within 9 months, we created a highly available real-time customer facing application for customer analytics. We currently have 500 different people doing their own analysis and projects with this data, spanning a total of 50 different use cases. This data, (consisting of over 40 million customer records with billions of transactions), provides our business new insights that were inaccessible before.
Our business moves quickly, with several products and 20 use cases currently in production. We currently have a customer data lake and a technical data lake. Having a platform with very different workloads has proven to be challenging.
Our success in generating value created such growth in terms of data, use cases, analysts and usage patterns that 3 years later we find issues with scalability in HDFS, Hive metastore and Hadoop operations and challenges with highly available architectures with Hbase, Flume and Kafka. Going forward we are exploring alternative architectures including a hybrid cloud model, and moving towards streaming.
Our goal with this session is to assist people in the early part of their journey by building a solid foundation. We hope that others can benefit from us sharing our experiences and lessons learned during our journey.
Speaker
Nicolette Bullivant, Head of Data Engineering at Santander UK Technology, Santander UK Technology
Oracle Exalytics - Tips and Experiences from the Field (Enkitec E4 Conference...Mark Rittman
Presentation by Rittman Mead's Mark Rittman and Stewart Bryson on our experiences 1-year on with Exalytics. Includes sections on aggregate caching and datamart loading into TT, use of Essbase as a TT alternative, and deployment patterns we see on client sites.
OBIEE, Endeca, Hadoop and ORE Development (on Exalytics) (ODTUG 2013)Mark Rittman
A presentation from ODTUG 2013 on tools other than OBIEE for Exalytics, focusing on analysis of non-traditional data via Endeca, "big data" via Hadoop and statistical analysis / predictive modeling through Oracle R Enterprise, and the benefits of running these tools on Oracle Exalytics
Inside Oracle Exalytics and Oracle TimesTen for Exalytics - Hotsos 2012Mark Rittman
Mark Rittman presented on Oracle Exalytics and Oracle TimesTen for Exalytics at the Hotsos Symposium 2012. He discussed (1) what Exalytics is as an in-memory appliance for Oracle Business Intelligence that combines specialized hardware and optimized software, (2) how it addresses performance issues for analytics workloads by caching data and aggregates in memory, and (3) its architecture which includes optimized versions of OBIEE and Essbase running on TimesTen for fast in-memory analytics.
Mark Rittman is an Oracle ACE Director and co-founder of Rittman Mead, a specialist Oracle BI consulting firm. He has over 15 years of experience with Oracle technologies including BI, OLAP, and the Oracle database. He is a regular speaker at Oracle OpenWorld and columnist for Oracle Magazine. He has authored two books on Oracle BI through Oracle Press. This document provides an introduction and overview of Oracle Business Intelligence including its semantic business model, interactive dashboards, and integration capabilities. Demonstrations are shown of the semantic model, dashboard creation, and integration with Oracle Fusion Middleware.
Ougn2013 high speed, in-memory big data analysis with oracle exalyticsMark Rittman
The document discusses Oracle Exalytics, a platform for high speed, big data analysis. Exalytics combines Oracle Business Intelligence software with specialized hardware to enable high-density visualization of large datasets and support of many concurrent users. It also integrates Oracle Essbase, Endeca, and Hadoop to provide additional analytic capabilities for both structured and unstructured data.
This document provides an overview of Oracle Endeca Information Discovery (EID) 2.3, a platform for data discovery and exploration. It discusses:
- Endeca's acquisition by Oracle and its focus on search and guided navigation.
- The key components of EID including the Endeca Server for storing hybrid search/analytic data, Oracle EID Integrator for ETL, and Oracle EID Studio for building interfaces.
- The development process including loading data into the Endeca Server using Integrator, configuring attributes and search, and building interfaces in Studio.
This document discusses Oracle's business intelligence and enterprise performance management products, with a focus on Oracle Exalytics. It notes that business analytics was the top priority for CIOs in 2012 according to Gartner. It describes Oracle Exalytics as optimized hardware for business intelligence and in-memory analytics. Examples are given of customers who saw significantly faster performance and reduced cycle times using Oracle Exalytics compared to alternative solutions.
IlOUG Tech Days 2016 - Unlock the Value in your Data Reservoir using Oracle B...Mark Rittman
Mark Rittman from Rittman Mead presented on Oracle Big Data Discovery. He discussed how many organizations are running big data initiatives involving loading large amounts of raw data into data lakes for analysis. Oracle Big Data Discovery provides a visual interface for exploring, analyzing, and transforming this raw data. It allows users to understand relationships in the data, perform enrichments, and prepare the data for use in tools like Oracle Business Intelligence.
Maxis is a data services company founded in 2010 that provides Maximo consulting, upgrades, and archival services using their Alchemize software. Alchemize started as a Maximo archive solution in 2015 and has since expanded to support full data migration, transformation, and analytics capabilities. Key features of Alchemize include its ability to archive and migrate data across different systems quickly and flexibly while maintaining data integrity. It is used by various organizations for projects such as compliance, system performance optimization, and application retirement.
Presentation by Mark Rittman, Technical Director, Rittman Mead, on ODI 11g features that support enterprise deployment and usage. Delivered at BIWA Summit 2013, January 2013.
ADV Slides: When and How Data Lakes Fit into a Modern Data ArchitectureDATAVERSITY
Whether to take data ingestion cycles off the ETL tool and the data warehouse or to facilitate competitive Data Science and building algorithms in the organization, the data lake – a place for unmodeled and vast data – will be provisioned widely in 2020.
Though it doesn’t have to be complicated, the data lake has a few key design points that are critical, and it does need to follow some principles for success. Avoid building the data swamp, but not the data lake! The tool ecosystem is building up around the data lake and soon many will have a robust lake and data warehouse. We will discuss policy to keep them straight, send data to its best platform, and keep users’ confidence up in their data platforms.
Data lakes will be built in cloud object storage. We’ll discuss the options there as well.
Get this data point for your data lake journey.
Deploying OBIEE11g in the Enterprise (UKOUG 2012)Mark Rittman
In this presentation, we provide tips, techniques and guidance on how to integrate OBIEE 11g into your enterprise's security, application server, management and diagnostics arrangements, and how OBIEE should be deployed for high availability, resilience and easy backup/recovery/cloning in an enterprise environment.
Pradeep Kumar Pandey has over 10 years of experience as a data/systems integration specialist and ETL expert. He has extensive experience designing and implementing data warehouses using tools like IBM DataStage, Informatica, Oracle OBIEE, and Oracle OBIA. He has led teams and taken on roles such as developer, technical lead, and team lead. Pradeep has worked on projects across various industries including telecom, financial services, HR, and retail.
How to Integrate OBIEE and Essbase / EPM Suite (OOW 2012)Mark Rittman
Oracle plans to integrate Oracle Essbase and the EPM product suite with Oracle Business Intelligence Enterprise Edition and Oracle Fusion Middleware. So with the latest release of Oracle Business Intelligence Enterprise Edition, 11.1.1.6, how do you connect Oracle Business Intelligence Enterprise Edition to your Oracle Essbase databases and how well does it handle Oracle Essbase features such as scenario and account dimensions, changing outlines, and unbalanced/parent-child hierarchies? How well do Oracle Business Intelligence Enterprise Edition’s ad hoc reporting tools handle Oracle Essbase hierarchies and member selections in the 11.1.1.6 release? Can we still embed Oracle Business Intelligence Enterprise Edition dashboards in Oracle Workspaces? Learn the answers in this session.
Powering a Startup with Apache Spark with Kevin KimSpark Summit
In Between (A mobile App for couples, downloaded 20M in Global), from daily batch for extracting metrics, analysis and dashboard. Spark is widely used by engineers and data analysts in Between, thanks to the performance and expendability of Spark, data operating has become extremely efficient. Entire team including Biz Dev, Global Operation, Designers are enjoying data results so Spark is empowering entire company for data driven operation and thinking. Kevin, Co-founder and Data Team leader of Between will be presenting how things are going in Between. Listeners will know how small and agile team is living with data (how we build organization, culture and technical base) after this presentation.
The document is a presentation on Oracle NoSQL Database that discusses its use cases, Oracle's NoSQL and big data strategy, technical features of Oracle NoSQL Database, and customer references. The presentation covers how Oracle NoSQL Database can be used for real-time event processing, sensor data acquisition, fraud detection, recommendations, and globally distributed databases. It also discusses Oracle's approach to integrating NoSQL, Hadoop, and relational databases. Customer references are provided for Airbus's use of Oracle NoSQL Database for flight test sensor data storage and analysis.
The document discusses the identity management system at the University of Edinburgh. It describes the current homegrown system, issues with scalability and cost, and an evaluation of open source and commercial identity management solutions. A blended solution was chosen using the open source Grouper system for group management and reusing existing Oracle and OpenLDAP components. This provided functionality needed while avoiding high licensing costs of a commercial solution.
This document summarizes Terry Bunio's presentation on breaking and fixing broken data. It begins by thanking sponsors and providing information about Terry Bunio and upcoming SQL events. It then discusses the three types of broken data: inconsistent, incoherent, and ineffectual data. For each type, it provides an example and suggestions on how to identify and fix the issues. It demonstrates how to use tools like Oracle Data Modeler, execution plans, SQL Profiler, and OStress to diagnose problems to make data more consistent, coherent and effective.
Enterprise Data World 2018 - Building Cloud Self-Service Analytical SolutionDmitry Anoshin
This session will cover building the modern Data Warehouse by migration from the traditional DW platform into the cloud, using Amazon Redshift and Cloud ETL Matillion in order to provide Self-Service BI for the business audience. This topic will cover the technical migration path of DW with PL/SQL ETL to the Amazon Redshift via Matillion ETL, with a detailed comparison of modern ETL tools. Moreover, this talk will be focusing on working backward through the process, i.e. starting from the business audience and their needs that drive changes in the old DW. Finally, this talk will cover the idea of self-service BI, and the author will share a step-by-step plan for building an efficient self-service environment using modern BI platform Tableau.
The Future of Analytics, Data Integration and BI on Big Data PlatformsMark Rittman
The document discusses the future of analytics, data integration, and business intelligence (BI) on big data platforms like Hadoop. It covers how BI has evolved from old-school data warehousing to enterprise BI tools to utilizing big data platforms. New technologies like Impala, Kudu, and dataflow pipelines have made Hadoop fast and suitable for analytics. Machine learning can be used for automatic schema discovery. Emerging open-source BI tools and platforms, along with notebooks, bring new approaches to BI. Hadoop has become the default platform and future for analytics.
Using Oracle Big Data Discovey as a Data Scientist's ToolkitMark Rittman
As delivered at Trivadis Tech Event 2016 - how Big Data Discovery along with Python and pySpark was used to build predictive analytics models against wearables and smart home data
SQL-on-Hadoop for Analytics + BI: What Are My Options, What's the Future?Mark Rittman
There are many options for providing SQL access over data in a Hadoop cluster, including proprietary vendor products along with open-source technologies such as Apache Hive, Cloudera Impala and Apache Drill; customers are using those to provide reporting over their Hadoop and relational data platforms, and looking to add capabilities such as calculation engines, data integration and federation along with in-memory caching to create complete analytic platforms. In this session we’ll look at the options that are available, compare database vendor solutions with their open-source alternative, and see how emerging vendors are going beyond simple SQL-on-Hadoop products to offer complete “data fabric” solutions that bring together old-world and new-world technologies and allow seamless offloading of archive data and compute work to lower-cost Hadoop platforms.
IlOUG Tech Days 2016 - Big Data for Oracle Developers - Towards Spark, Real-T...Mark Rittman
Mark Rittman, CTO of Rittman Mead, gave a keynote presentation on big data for Oracle developers and DBAs with a focus on Apache Spark, real-time analytics, and predictive analytics. He discussed how Hadoop can provide flexible, cheap storage for logs, feeds, and social data. He also explained several Hadoop processing frameworks like Apache Spark, Apache Tez, Cloudera Impala, and Apache Drill that provide faster alternatives to traditional MapReduce processing.
Enkitec E4 Barcelona : SQL and Data Integration Futures on Hadoop : Mark Rittman
Mark Rittman gave a presentation on the future of analytics on Oracle Big Data Appliance. He discussed how Hadoop has enabled highly scalable and affordable cluster computing using technologies like MapReduce, Hive, Impala, and Parquet. Rittman also talked about how these technologies have improved query performance and made Hadoop suitable for both batch and interactive/ad-hoc querying of large datasets.
Oracle BI Hybrid BI : Mode 1 + Mode 2, Cloud + On-Premise Business AnalyticsMark Rittman
Mark Rittman, founder of Rittman Mead, discusses Oracle's approach to hybrid BI deployments and how it aligns with Gartner's vision of a modern BI platform. He explains how Oracle BI 12c supports both traditional top-down modeling and bottom-up data discovery. It also enables deploying components on-premises or in the cloud for flexibility. Rittman believes the future is bi-modal, with IT enabling self-service analytics alongside centralized governance.
Deploying Full Oracle BI Platforms to Oracle Cloud - OOW2015Mark Rittman
- Mark Rittman presented on deploying full OBIEE systems to Oracle Cloud. This involves migrating the data warehouse to Oracle Database Cloud Service, updating the RPD to connect to the cloud database, and uploading the RPD to Oracle BI Cloud Service. Using the wider Oracle PaaS ecosystem allows hosting a full BI platform in the cloud.
Delivering the Data Factory, Data Reservoir and a Scalable Oracle Big Data Ar...Mark Rittman
Presentation from the Rittman Mead BI Forum 2015 masterclass, pt.2 of a two-part session that also covered creating the Discovery Lab. Goes through setting up Flume log + twitter feeds into CDH5 Hadoop using ODI12c Advanced Big Data Option, then looks at the use of OBIEE11g with Hive, Impala and Big Data SQL before finally using Oracle Big Data Discovery for faceted search and data mashup on-top of Hadoop
OBIEE11g Seminar by Mark Rittman for OU Expert Summit, Dubai 2015Mark Rittman
Slides from a two-day OBIEE11g seminar in Dubai, February 2015, at the Oracle University Expert Summit. Covers the following topics:
1. OBIEE 11g Overview & New Features
2. Adding Exalytics and In-Memory Analytics to OBIEE 11g
3. Source Control and Concurrent Development for OBIEE
4. No Silver Bullets - OBIEE 11g Performance in the Real World
5. Oracle BI Cloud Service Overview, Tips and Techniques
6. Moving to Oracle BI Applications 11g + ODI
7. Oracle Essbase and Oracle BI EE 11g Integration Tips and Techniques
8. OBIEE 11g and Predictive Analytics, Hadoop & Big Data
BIWA2015 - Bringing Oracle Big Data SQL to OBIEE and ODIMark Rittman
The document discusses Oracle's Big Data SQL, which brings Oracle SQL capabilities to Hadoop data stored in Hive tables. It allows querying Hive data using standard SQL from Oracle Database and viewing Hive metadata in Oracle data dictionary tables. Big Data SQL leverages the Hive metastore and uses direct reads and SmartScan to optimize queries against HDFS and Hive data. This provides a unified SQL interface and optimized query processing for both Oracle and Hadoop data.
UKOUG Tech'14 Super Sunday : Deep-Dive into Big Data ETL with ODI12cMark Rittman
This document discusses using Hadoop and Hive for ETL work. It provides an overview of using Hadoop for distributed processing and storage of large datasets. It describes how Hive provides a SQL interface for querying data stored in Hadoop and how various Apache tools can be used to load, transform and store data in Hadoop. Examples of using Hive to view table metadata and run queries are also presented.
Part 4 - Hadoop Data Output and Reporting using OBIEE11gMark Rittman
Delivered as a one-day seminar at the SIOUG and HROUG Oracle User Group Conferences, October 2014.
Once insights and analysis have been produced within your Hadoop cluster by analysts and technical staff, it’s usually the case that you want to share the output with a wider audience in the organisation. Oracle Business Intelligence has connectivity to Hadoop through Apache Hive compatibility, and other Oracle tools such as Oracle Big Data Discovery and Big Data SQL can be used to visualise and publish Hadoop data. In this final session we’ll look at what’s involved in connecting these tools to your Hadoop environment, and also consider where data is optimally located when large amounts of Hadoop data need to be analysed alongside more traditional data warehouse datasets
Generative Artificial Intelligence (GenAI) in BusinessDr. Tathagat Varma
My talk for the Indian School of Business (ISB) Emerging Leaders Program Cohort 9. In this talk, I discussed key issues around adoption of GenAI in business - benefits, opportunities and limitations. I also discussed how my research on Theory of Cognitive Chasms helps address some of these issues
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptxJustin Reock
Building 10x Organizations with Modern Productivity Metrics
10x developers may be a myth, but 10x organizations are very real, as proven by the influential study performed in the 1980s, ‘The Coding War Games.’
Right now, here in early 2025, we seem to be experiencing YAPP (Yet Another Productivity Philosophy), and that philosophy is converging on developer experience. It seems that with every new method we invent for the delivery of products, whether physical or virtual, we reinvent productivity philosophies to go alongside them.
But which of these approaches actually work? DORA? SPACE? DevEx? What should we invest in and create urgency behind today, so that we don’t find ourselves having the same discussion again in a decade?
Technology Trends in 2025: AI and Big Data AnalyticsInData Labs
At InData Labs, we have been keeping an ear to the ground, looking out for AI-enabled digital transformation trends coming our way in 2025. Our report will provide a look into the technology landscape of the future, including:
-Artificial Intelligence Market Overview
-Strategies for AI Adoption in 2025
-Anticipated drivers of AI adoption and transformative technologies
-Benefits of AI and Big data for your business
-Tips on how to prepare your business for innovation
-AI and data privacy: Strategies for securing data privacy in AI models, etc.
Download your free copy nowand implement the key findings to improve your business.
Mobile App Development Company in Saudi ArabiaSteve Jonas
EmizenTech is a globally recognized software development company, proudly serving businesses since 2013. With over 11+ years of industry experience and a team of 200+ skilled professionals, we have successfully delivered 1200+ projects across various sectors. As a leading Mobile App Development Company In Saudi Arabia we offer end-to-end solutions for iOS, Android, and cross-platform applications. Our apps are known for their user-friendly interfaces, scalability, high performance, and strong security features. We tailor each mobile application to meet the unique needs of different industries, ensuring a seamless user experience. EmizenTech is committed to turning your vision into a powerful digital product that drives growth, innovation, and long-term success in the competitive mobile landscape of Saudi Arabia.
Dev Dives: Automate and orchestrate your processes with UiPath MaestroUiPathCommunity
This session is designed to equip developers with the skills needed to build mission-critical, end-to-end processes that seamlessly orchestrate agents, people, and robots.
📕 Here's what you can expect:
- Modeling: Build end-to-end processes using BPMN.
- Implementing: Integrate agentic tasks, RPA, APIs, and advanced decisioning into processes.
- Operating: Control process instances with rewind, replay, pause, and stop functions.
- Monitoring: Use dashboards and embedded analytics for real-time insights into process instances.
This webinar is a must-attend for developers looking to enhance their agentic automation skills and orchestrate robust, mission-critical processes.
👨🏫 Speaker:
Andrei Vintila, Principal Product Manager @UiPath
This session streamed live on April 29, 2025, 16:00 CET.
Check out all our upcoming Dev Dives sessions at https://community.uipath.com/dev-dives-automation-developer-2025/.
This is the keynote of the Into the Box conference, highlighting the release of the BoxLang JVM language, its key enhancements, and its vision for the future.
Linux Support for SMARC: How Toradex Empowers Embedded DevelopersToradex
Toradex brings robust Linux support to SMARC (Smart Mobility Architecture), ensuring high performance and long-term reliability for embedded applications. Here’s how:
• Optimized Torizon OS & Yocto Support – Toradex provides Torizon OS, a Debian-based easy-to-use platform, and Yocto BSPs for customized Linux images on SMARC modules.
• Seamless Integration with i.MX 8M Plus and i.MX 95 – Toradex SMARC solutions leverage NXP’s i.MX 8 M Plus and i.MX 95 SoCs, delivering power efficiency and AI-ready performance.
• Secure and Reliable – With Secure Boot, over-the-air (OTA) updates, and LTS kernel support, Toradex ensures industrial-grade security and longevity.
• Containerized Workflows for AI & IoT – Support for Docker, ROS, and real-time Linux enables scalable AI, ML, and IoT applications.
• Strong Ecosystem & Developer Support – Toradex offers comprehensive documentation, developer tools, and dedicated support, accelerating time-to-market.
With Toradex’s Linux support for SMARC, developers get a scalable, secure, and high-performance solution for industrial, medical, and AI-driven applications.
Do you have a specific project or application in mind where you're considering SMARC? We can help with Free Compatibility Check and help you with quick time-to-market
For more information: https://www.toradex.com/computer-on-modules/smarc-arm-family
Quantum Computing Quick Research Guide by Arthur MorganArthur Morgan
This is a Quick Research Guide (QRG).
QRGs include the following:
- A brief, high-level overview of the QRG topic.
- A milestone timeline for the QRG topic.
- Links to various free online resource materials to provide a deeper dive into the QRG topic.
- Conclusion and a recommendation for at least two books available in the SJPL system on the QRG topic.
QRGs planned for the series:
- Artificial Intelligence QRG
- Quantum Computing QRG
- Big Data Analytics QRG
- Spacecraft Guidance, Navigation & Control QRG (coming 2026)
- UK Home Computing & The Birth of ARM QRG (coming 2027)
Any questions or comments?
- Please contact Arthur Morgan at [email protected].
100% human made.
UiPath Community Berlin: Orchestrator API, Swagger, and Test Manager APIUiPathCommunity
Join this UiPath Community Berlin meetup to explore the Orchestrator API, Swagger interface, and the Test Manager API. Learn how to leverage these tools to streamline automation, enhance testing, and integrate more efficiently with UiPath. Perfect for developers, testers, and automation enthusiasts!
📕 Agenda
Welcome & Introductions
Orchestrator API Overview
Exploring the Swagger Interface
Test Manager API Highlights
Streamlining Automation & Testing with APIs (Demo)
Q&A and Open Discussion
Perfect for developers, testers, and automation enthusiasts!
👉 Join our UiPath Community Berlin chapter: https://community.uipath.com/berlin/
This session streamed live on April 29, 2025, 18:00 CET.
Check out all our upcoming UiPath Community sessions at https://community.uipath.com/events/.
TrsLabs - Fintech Product & Business ConsultingTrs Labs
Hybrid Growth Mandate Model with TrsLabs
Strategic Investments, Inorganic Growth, Business Model Pivoting are critical activities that business don't do/change everyday. In cases like this, it may benefit your business to choose a temporary external consultant.
An unbiased plan driven by clearcut deliverables, market dynamics and without the influence of your internal office equations empower business leaders to make right choices.
Getting things done within a budget within a timeframe is key to Growing Business - No matter whether you are a start-up or a big company
Talk to us & Unlock the competitive advantage
AI Changes Everything – Talk at Cardiff Metropolitan University, 29th April 2...Alan Dix
Talk at the final event of Data Fusion Dynamics: A Collaborative UK-Saudi Initiative in Cybersecurity and Artificial Intelligence funded by the British Council UK-Saudi Challenge Fund 2024, Cardiff Metropolitan University, 29th April 2025
https://alandix.com/academic/talks/CMet2025-AI-Changes-Everything/
Is AI just another technology, or does it fundamentally change the way we live and think?
Every technology has a direct impact with micro-ethical consequences, some good, some bad. However more profound are the ways in which some technologies reshape the very fabric of society with macro-ethical impacts. The invention of the stirrup revolutionised mounted combat, but as a side effect gave rise to the feudal system, which still shapes politics today. The internal combustion engine offers personal freedom and creates pollution, but has also transformed the nature of urban planning and international trade. When we look at AI the micro-ethical issues, such as bias, are most obvious, but the macro-ethical challenges may be greater.
At a micro-ethical level AI has the potential to deepen social, ethnic and gender bias, issues I have warned about since the early 1990s! It is also being used increasingly on the battlefield. However, it also offers amazing opportunities in health and educations, as the recent Nobel prizes for the developers of AlphaFold illustrate. More radically, the need to encode ethics acts as a mirror to surface essential ethical problems and conflicts.
At the macro-ethical level, by the early 2000s digital technology had already begun to undermine sovereignty (e.g. gambling), market economics (through network effects and emergent monopolies), and the very meaning of money. Modern AI is the child of big data, big computation and ultimately big business, intensifying the inherent tendency of digital technology to concentrate power. AI is already unravelling the fundamentals of the social, political and economic world around us, but this is a world that needs radical reimagining to overcome the global environmental and human challenges that confront us. Our challenge is whether to let the threads fall as they may, or to use them to weave a better future.
Increasing Retail Store Efficiency How can Planograms Save Time and Money.pptxAnoop Ashok
In today's fast-paced retail environment, efficiency is key. Every minute counts, and every penny matters. One tool that can significantly boost your store's efficiency is a well-executed planogram. These visual merchandising blueprints not only enhance store layouts but also save time and money in the process.
TrustArc Webinar: Consumer Expectations vs Corporate Realities on Data Broker...TrustArc
Most consumers believe they’re making informed decisions about their personal data—adjusting privacy settings, blocking trackers, and opting out where they can. However, our new research reveals that while awareness is high, taking meaningful action is still lacking. On the corporate side, many organizations report strong policies for managing third-party data and consumer consent yet fall short when it comes to consistency, accountability and transparency.
This session will explore the research findings from TrustArc’s Privacy Pulse Survey, examining consumer attitudes toward personal data collection and practical suggestions for corporate practices around purchasing third-party data.
Attendees will learn:
- Consumer awareness around data brokers and what consumers are doing to limit data collection
- How businesses assess third-party vendors and their consent management operations
- Where business preparedness needs improvement
- What these trends mean for the future of privacy governance and public trust
This discussion is essential for privacy, risk, and compliance professionals who want to ground their strategies in current data and prepare for what’s next in the privacy landscape.
TrustArc Webinar: Consumer Expectations vs Corporate Realities on Data Broker...TrustArc
In-Memory Oracle BI Applications (UKOUG Analytics Event, July 2013)
1. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
In-Memory Analytics with Oracle BI Apps and Oracle Exalytics
UKOUG Analytics Event, London, July 2013
Mark Rittman, Technical Director, Rittman Mead
1Friday, 5 July 13
2. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
About the Speaker
•Mark Rittman, Co-Founder of Rittman Mead
•Oracle ACE Director, specialising in Oracle BI&DW
•14 Years Experience with Oracle Technology
•Regular columnist for Oracle Magazine
•Author of two Oracle Press Oracle BI books
•Oracle Business Intelligence Developers Guide
•Oracle Exalytics Revealed
•Writer for Rittman Mead Blog :
http://www.rittmanmead.com/blog
•Email : [email protected]
•Twitter : @markrittman
2Friday, 5 July 13
3. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
About Rittman Mead
•Oracle BI and DW platinum partner
•World leading specialist partner for technical excellence, solutions delivery and innovation in Oracle BI
•Approximately 50 consultants worldwide
•All expert in Oracle BI and DW
•Offices in US (Atlanta), Europe, Australia and India
•Skills in broad range of supporting Oracle tools:
‣ OBIEE
‣ OBIA
‣ ODIEE
‣ Essbase, Oracle OLAP
‣ GoldenGate
‣ Exadata
‣ Endeca
3Friday, 5 July 13
4. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Oracle Business Intelligence 11g
•Oracle’s business intelligence platform, 11.1.1.7 release came out in April 2013
•Fourth major release of OBIEE 11g, with many new features + updated look and feel
•Enterprise BI platform centered around the Common Enterprise Semantic Model (RPD)
•Mobile BI apps, MS Office integration, ad-hoc,
dashboard and published reporting
•Built around Oracle Fusion Middleware
•Deployable on Windows, Unix, Linux
•Accessing a range of enterprise data sources
‣ Oracle and other RDBMSs
‣ Essbase and other OLAP servers
‣ Files, XML, web services
‣ ADF and SOA sources
‣ TimesTen in-memory database
4Friday, 5 July 13
5. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Oracle BI Applications
•Packaged version of OBIEE that includes a data warehouse, and ETL mappings,
from E-Business Suite, Siebel, SAP and Peoplesoft
•Covers areas such as Financial Analytics, HR Analytics, Sales Analytics etc
•Built on the same technology as OBIEE 11g, plus ETL and administration tools
5Friday, 5 July 13
6. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Pre-Defined ETL Routines from Oracle EBS, Peoplesoft, Siebel, JDE, SAP
•Integrated, conformed dimensional data warehouse
•Deployable on Oracle, MS SQL, IBM DB/2 and Teradata
•Uses Informatica PowerCenter for ETL, or now ODI11g
•Staging tables and presentation tables
•Allows modular deployment
•Lowest grain of information
•Prebuilt aggregates
•History tracking
•Indexing
6Friday, 5 July 13
7. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
BI Apps Data Warehouse Limitations
•Designed for “lowest common denominator” DB features
‣ No “out of the box” partitioning, MVs, compression optimization for PQ
•Based on traditional disk-based RBDMS technology
•Can often lead to slow reports, dashboards, limiting user acceptance
•Common issue - what can we do about it?
7Friday, 5 July 13
8. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Oracle Exalytics : First In-Memory Engineered System for Analytics
• Relational, Multi-Dimensional and Unstructured data analysis available as a single engineered system
• Combination of in-memory hardware and optimized software versions
• Supports the Exadata and Big Data Appliance data management systems
Exalytics
In-Memory
Machine
Spans Relational, Multi-Dimensional, and Unstructured analysis,
combined with Financial & Operational Planning
‣ In-Memory Optimized Hardware
‣ In-Memory Oracle BI, TimesTen, Essbase, and Endeca
‣ Many In-Memory Software Innovations
Tightly-Integrated with Exadata, and
Big Data Appliance
8Friday, 5 July 13
9. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Oracle Exalytics Benefits vs. Standard Hardware + Software
•Single supported stack of hardware + software : patching synchronized and tested across all components
•OBIEE, Essbase, TimesTen etc optimizations that are only available when deployed on Exalytics hardware
•Optimal selection of CPUs, RAM (DRAM), network connectors for a BI application tier
•Automatic in-memory caching of commonly-used aggregates - no manual tuning and selection
•Future platform for all Oracle BI products - EPM Suite, BI, Endeca, BI Apps
9Friday, 5 July 13
10. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Exalytics as the Exa-Machine for OBIEE
•Runs the BI layer on a high-performance, multi-core, 1TB server
•In-memory cache used to accelerate the BI part of the stack
•If Exadata addresses 80% of the query performance, Exalytics addresses
the remaining 20%
‣ Consistent response times for queries
‣ In-memory caching of aggregates
‣ 40 cores for high concurrency
‣ Re-engineered BI and OLAP software
that assumes 40 cores and 1TB RAM
ERP/Apps DW
Oracle BI
In-Memory DB/Cache
10Friday, 5 July 13
11. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Exalytics Under the Covers - How Does it Work?
•Exalytics brings together different technologies, which are still standalone products in their own right
•To harmonise and optimise their use within Exalytics, it utilises the following techniques:
‣ In-Memory Adaptive Data Mart - Using Oracle TimesTen for Exalytics, an in-memory RDBMS
‣ In-Memory Intelligent Result Cache
‣ In-Memory Cubes
•Some of these are genuine "secret sauce"
•New functionality and algorithms
•You can only get them through licensing Exalytics
•Others are descriptions of DW/BI strategies, or existing product functionality, extended to take advantage of
the capacity for processing in memory that Exalytics has
11Friday, 5 July 13
12. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
In-Memory Adaptive Data Mart
•Commonly-used aggregates are copied into Oracle TimesTen for Exalytics
•Past query patterns are analyzed and suitable aggregates recommended
•Oracle BI Server then uses these aggregates to make queries run faster
•Aggregates change over time in response to
changes in query patterns
•Tools are provided for managing and populating these aggregates
TimesTen BI Server
Exalytics
Aggregates
Data Warehouse
Detail-level
Data
12Friday, 5 July 13
13. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Oracle TimesTen for Exalytics
•New version of TimesTen specifically for Exalytics (and only available with Exalytics)
•Support for analytic functions
‣ Perform all the processing at source
‣ Combine with being in-memory = should be very fast
•Column compression
‣ Whitepaper cites 5x
‣ Given the hardware capacity, we could seriously contemplate loading the whole Data Warehouse into
memory
‣ Opens up lots of interesting design potential
•We can load aggregates into TimesTen, leave base data at source, and use OBIEE’s Vertical Federation
capability to seamlessly report across both
‣ All hidden from the end-user, all they will know is that their reports run fast!
13Friday, 5 July 13
14. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
TimesTen and OBIEE Architecture
•Oracle BI Server communicates with TimesTen through TimesTen Client
•Summary Advisor, and nqcmd use Oracle BI Server to access TimesTen
•Typical single TimesTen database per Exalyics machine
‣ Max TimesTen database size around 300MB
- Due to need to set aside equal
Temp size for the Perm size selected
•Clustered Exalytics boxes can be daisy-chained
together using InfinBand connections
‣ For HA scenarios, does not increase
available RAM
‣ Summary advisor scripts write to both TimesTen
databases, replicating aggregates
‣ TimesTen databases can be “wired together”
for failover/HA purposes
TimesTen
Memory-Resident
Database
Checkpoint
Files
Log
Files
ODBC
Oracle BI
Server
nqcmdSummary
Advisor
14Friday, 5 July 13
15. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Summary Advisor for Aggregate Recommendation & Creation
•Utility within Oracle BI Administrator tool that recommends aggregates
•Bases recommendations on usage tracking and summary statistics data
•Captured based on past activity
•Runs an iterative algorithm that searches,
each iteration, for the best aggregate
•Could we use this to cache commonly-used BI Apps
aggregations in TimesTen, automatically based on usage patterns?
15Friday, 5 July 13
16. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Alternative Approach - Copy “Hot Data” into TimesTen for Exalytics
•Standard approach is to store aggregates in the TimesTen datamart
‣ Aggregated by the source DB, aggregates then cached in TT database
•Other approaches could be used, however
‣ Store whole detail-level dataset in the TT database
‣ Store just recent detail-level data in TT, and use OBIEE’s fragmentation feature
‣ Store aggregate layer from BI Apps DW entirely in TimesTen
•Would this be an option that we could use with BI Apps datasets?
16Friday, 5 July 13
17. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Another Option - Oracle In-Memory Database Cache (IMDB)
•Automatically replicate “hot” transactional data from Oracle BI Apps DW tables into TimesTen for Exalytics
•Use OBIEE fragmentation to enable automatic navigation between sources
•Aggregation performed by both TimesTen,
and by source DB (as appropriate)
•However - fairly intrusive approach, Oracle-only,
probably not attractive to most
BI Apps customers and DBAs
17Friday, 5 July 13
18. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Running BI Apps In-Memory - Is it Possible and Practical?
•Can we use the Summary Advisor to automatically cache commonly-used aggregates in-memory?
‣ Similar to regular OBIEE caching, relies on query repeatability + use of aggregation
•Could we copy all, or part, of the BI Apps data warehouse directly into TimesTen?
•How would we update the RPD to point to the in-memory tables?
•How fast would TimesTen be to load, and to query, vs. Oracle/SQL Server/DB2 etc?
•Here’s our thoughts and R&D to date....
18Friday, 5 July 13
19. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Using the Summary Advisor on BI Apps 7.9.6.4 Step 1 : Start Advisor
•Server has to be an Exalytics server, in this example is patched-up to 11.1.1.6.9
•Workstation has the 11.1.1.6.9 BI Administration tool installed
•Select Tools > Utilities, then Oracle BI Summary Advisor from utility list
1
2
19Friday, 5 July 13
20. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Using the Summary Advisor on BI Apps 7.9.6.4 Step 2 : Scope Source Queries
•By default, all queries registered in the usage tracking and summary statistics tables are in-scope
•Refine the recommendations by limiting timeframe, and setting minimum accumulated time threshold
•Still an opportunity later on to pick and choose from recommended aggregates
•Once selected, then select the TimesTen connection pool and database as the aggregate table target
3
4
20Friday, 5 July 13
21. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Using the Summary Advisor on BI Apps 7.9.6.4 Step 3 : Create Aggregates
•Summary Advisor then recommends a set of “candidate” aggregates, which you can choose to implement
•Select all, none or some of the recommended aggregates
•Then run the resulting logical SQL script using the nqcmd utility
•Note - may need to clean-up BI Apps DW data to remove duplicates etc before script completes OK
21Friday, 5 July 13
22. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Using the Summary Advisor on BI Apps 7.9.6.4 Step 4 : Review RPD and Data
•Aggregate Persistence process called by the “create aggregates” process also maps tables in RPD
•Physical layer contains entries for the TimesTen tables
•Business Model and Mapping later contains vertically-federated LTSs for the new TT tables
22Friday, 5 July 13
23. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Summary Advisor for BI Apps : Pros and Cons
•Pros
‣ Automatically analyzes query patterns and recommends aggregates to accelerate them
‣ Handles the registering of the TimesTen tables in the RPD, including mapping back into business model
‣ Supports any data source that the BI Server supports
•Cons
‣ Queries have to have run before they’ll be considered for loading
into TimesTen for Exalytics
‣ Relies on subsequent queries being able to use those aggregates
‣ Could get unwieldy if many aggregates are registered in the RPD
‣ Summary Advisor process does not automatically clear down
tables that don’t feature in future recommendations
‣ Inefficient refresh process, unless you use a process such as
http://www.rittmanmead.com/2013/04/incremental
-refresh-of-exalytics-aggregates-using-
native-bi-server-capabilities/
23Friday, 5 July 13
24. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Alternative Approach - Copy “Hot Data” into TimesTen for Exalytics
•Standard approach is to store aggregates in the TimesTen datamart
‣ Aggregated by the source DB, aggregates then cached in TT database
•Other approaches could be used, however
‣ Store whole detail-level dataset in the TT database
‣ Store just recent detail-level data in TT, and use OBIEE’s fragmentation feature
‣ Store aggregate layer from BI Apps DW entirely in TimesTen
•Would this be an option that we could use with BI Apps datasets?
24Friday, 5 July 13
25. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Loading data directly from Oracle into TimesTen with ttLoadFromOracle
•The procedure ttLoadFromOracle uses OCI to load data directly
from Oracle into an existing TimesTen table
•Specify a whole table (SELECT *), or part (SELECT ... WHERE)
•Target table must existing on TimesTen already
‣ Create it automagically using ttTableSchemaFromOraQueryGet
or createandloadfromoraquery
‣ However both these use Oracle data types and no compression,
so size in memory is going to be greater
REVENUE_F_TS
REVENUE_F_LARGE
ttIsql --ConnStr
"DSN=BISAMPLE_TT;UID=SH;PWD=SH;OracleNetServiceName=
orcl;OraclePWD=SH"
Command> call ttLoadFromOracle('A_TEST',
'REVENUE_F_TS', 'SELECT SHIPTO_ADDR_KEY, OFFICE_KEY,
EMPL_KEY, PROD_KEY, ORDER_KEY, REVENUE, UNITS,
DISCNT_VALUE, BILL_MTH_KEY, BILL_QTR_KEY,
BILL_DAY_DT, ORDER_DAY_DT, PAID_DAY_DT, DISCNT_RATE,
ORDER_STATUS, CURRENCY, ORDER_TYPE, CUST_KEY,
SHIP_DAY_DT, COST_FIXED, COST_VARIABLE,
SRC_ORDER_NUMBER, ORDER_NUMBER FROM
BISAMPLE.SAMP_REVENUE_F_LARGE WHERE
BILL_MTH_KEY=201012');
< 7750 >
1 row found.
-- this has loaded 7750 rows for a given month
TimesTen is
loaded from the
results of a query
on Oracle
ttLoadFromOracle
25Friday, 5 July 13
26. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Generating TimesTen load DDL and scripts with ttImportFromOracle
•For ttLoadFromOracle to work, the target table must exist
•The utility ttImportFromOracle is useful here.
‣ It can map Oracle data types to optimal TimesTen ones
‣ Optionally, it can aggressively limit column sizes based on data to reduce TimesTen footprint
‣ It can evaluate compression effectiveness and apply it only where most useful
‣ Given a set of tables, it will generate:
- TimesTen DDL for requires schemas/tables/indexes
- A script to load all the tables into TimesTen in parallel (ttPDL.sh)
CreateIndexes.sql
ttImportFromOracle
REVENUE_F
CreateTables.sql
CreateUsers.sql
DropIndexes.sql
DropTables.sql
LoadData.sql TableList.txt
ttPDL.sh ttSizing.sh
UpdateStats.sql
26Friday, 5 July 13
27. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Generating TimesTen load DDL and scripts with ttImportFromOracle
•ttImportFromOracle is new in TimesTen 11.2.2.5
‣ Not an official production utility - best efforts support only
‣ But it’s only generating scripts, which contain standard (supported) TimesTen functionality
•The executable is located in $TT_HOME/support
•It uses OCI, so make sure LD_LIBRARY_PATH is set to include Oracle DB lib
‣ export LD_LIBRARY_PATH=$ORACLE_HOME/lib
•Feature-rich syntax, but at its simplest can just be invoked for a single table, with compression:
$ ttImportFromOracle -oraconn SH/SH@orcl -tables REVENUE_F_TS -compression 1
Beginning processing
Resolving any tablename wildcards
Eliminating any duplicate tables
Getting metadata from source
Generating database user list
Assigning TimesTen datatypes
Analyzing source tables
Analyzing table 'SH.REVENUE_F_TS' ...
Estimating table sizes
Evaluating parallel data load
Generating output files
Finished processing
27Friday, 5 July 13
28. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Generating TimesTen load DDL and scripts with ttImportFromOracle
•ttImportFromOracle writes a set of scripts that are subsequently executed to :
‣ Create the target tables and indexes on TimesTen, using optimised data
types and compression
‣ Load the target tables on TimesTen, still via ttLoadFromOracle
REVENUE_F
ttPDL.sh
REVENUE_F_TS
ttLoadFromOracle
REVENUE_F_TS
CREATE TABLE ...CreateTables.sql
28Friday, 5 July 13
29. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Comparing Oracle and TimesTen as data sources
•Our testing has shown that in general,
‣ For base level data, Oracle outperforms TimesTen
‣ For aggregated data, TimesTen outperforms Oracle
•Therefore entire lift + shift of OBIA Data Warehouse into TimesTen is possibly not going to give optimal
response times
-- same query over OBIA
-- 0.68 seconds
Select * from samp_revenue_f_large f, samp_customers_d
cd, samp_addresses_d ad, samp_products_d pd
where cd.cust_key = f.Cust_key and cd.address_key =
ad.address_key
and pd.prod_key = f.prod_key
and ad.city = 'San Francisco'
and pd.type = 'Cell Phones'
and f.bill_mth_key = 201012
;
-- using full sized fact table (native data types)
-- 0.78 seconds query time
Select * from revenue_f_native f, customers_d cd,
addresses_cd ad, products_d pd
where cd.cust_key = f.Cust_key and cd.address_key =
ad.address_key
and pd.prod_key = f.prod_key
and ad.city = 'San Francisco'
and pd.type = 'Cell Phones'
and f.bill_mth_key=201012
;
0.68 seconds 0.78 seconds
29Friday, 5 July 13
30. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Comparing Oracle and TimesTen as data sources
•One option would be to put just “Hot Data” (eg current month) into
TimesTen, and then update the RPD to use fragmentation
‣ This has an overhead in terms of RPD updates (and support - added
complexity), as well as an additional “ETL” process to manage
-- using full sized fact table (native data types)
-- 0.78 seconds query time
Select * from revenue_f_native f, customers_d cd,
addresses_cd ad, products_d pd
where cd.cust_key = f.Cust_key and cd.address_key =
ad.address_key
and pd.prod_key = f.prod_key
and ad.city = 'San Francisco'
and pd.type = 'Cell Phones'
and f.bill_mth_key=201012; 0.78 seconds
-- using Time slice table
-- 0.25 seconds
Select * from revenue_f_ts f, customers_d cd, addresses_cd
ad, products_d pd
where cd.cust_key = f.Cust_key and cd.address_key =
ad.address_key
and pd.prod_key = f.prod_key
and ad.city = 'San Francisco'
and pd.type = 'Cell Phones'
and f.bill_mth_key=201012; 0.25 seconds
30Friday, 5 July 13
31. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Conclusions
•The scale of most BI Apps implementations means that query performance can be an issue
•Exalytics’ TimesTen In-Memory Database could be a potential solution to this issue
•Several approaches to putting all or part of the BI Apps DW into TimesTen / In-Memory
•Copying all or sections of the BI Apps DW into TimesTen is a potential approach
‣ But current version of TimesTen best suited to smaller tables and datasets
•Exalytics’ Summary Advisor now works with Oracle BI Apps 7.9.6.4
•Automatically detects and recommends suitable aggregates, builds and maps into RPD
‣ Though custom solutions are probably more efficient for their later incremental refresh
•A work in progress - speak to Rittman Mead for more details on how this can work
•Offers the potential of “speed-of-thought” business analytics dashboards, with minimum additional work
31Friday, 5 July 13
32. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
Thank You for Attending!
• Thank you for attending this presentation, and more information can be found at http://www.rittmanmead.com
• Contact us at [email protected] or [email protected]
• Look out for our book, “Oracle Business Intelligence Developers Guide” out now!
• Follow-us on Twitter (@rittmanmead) or Facebook (facebook.com/rittmanmead)
32Friday, 5 July 13
33. T : +44 (0) 8446 697 995 E : [email protected] W: www.rittmanmead.com
In-Memory Analytics with Oracle BI Apps and Oracle Exalytics
UKOUG Analytics Event, London, July 2013
Mark Rittman, Technical Director, Rittman Mead
33Friday, 5 July 13