If you've invested in OBIEE and want to start exploring the use of Big Data technology, this presentation talks about how and why you might want to use OBIEE as the common visualization layer across both.
Offload, Transform, and Present - the New World of Data IntegrationMichael Rainey
How much time and effort (and budget) do organizations spend moving data around the enterprise? Unfortunately, quite a lot. These days, ETL developers are tasked with performing the Extract (E) and Load (L), and spending less time on their craft, building Transformations (T). This changes in the new world of data integration. By offloading data from the RDBMS to Hadoop, with the ability to present it back to the relational database, data can be seamlessly integrated between different source and target systems. Transformations occur on data offloaded to Hadoop, using the latest ETL technologies, or in the target database, with a standard ETL-on-RDBMS tool. In this session, we’ll discuss how the new world of data integration will provide focus on transforming data into insightful information by simplifying the data movement process.
Presented at Enkitec E4 2017.
A Walk Through the Kimball ETL Subsystems with Oracle Data Integration - Coll...Michael Rainey
Big Data integration is an excellent feature in the Oracle Data Integration product suite (Oracle Data Integrator, GoldenGate, & Enterprise Data Quality). But not all analytics require big data technologies, such as labor cost, revenue, or expense reporting. Ralph Kimball, an original architect of the dimensional model in data warehousing, spent much of his career working to build an enterprise data warehouse methodology that can meet these reporting needs. His book, "The Data Warehouse ETL Toolkit", is a guide for many ETL developers. This session will walk you through his ETL Subsystem categories; Extracting, Cleaning & Conforming, Delivering, and Managing, describing how the Oracle Data Integration products are perfectly suited for the Kimball approach.
Presented at Collaborate16 in Las Vegas.
A Walk Through the Kimball ETL Subsystems with Oracle Data IntegrationMichael Rainey
Big Data integration is an excellent feature in the Oracle Data Integration product suite (Oracle Data Integrator, GoldenGate, & Enterprise Data Quality). But not all analytics require big data technologies, such as labor cost, revenue, or expense reporting. Ralph Kimball, an original architect of the dimensional model in data warehousing, spent much of his career working to build an enterprise data warehouse methodology that can meet these reporting needs. His book, "The Data Warehouse ETL Toolkit", is a guide for many ETL developers. This session will walk you through his ETL Subsystem categories; Extracting, Cleaning & Conforming, Delivering, and Managing, describing how the Oracle Data Integration products are perfectly suited for the Kimball approach.
Presented at Oracle OpenWorld 2015 & BIWA Summit 2016.
Practical Tips for Oracle Business Intelligence Applications 11g ImplementationsMichael Rainey
The document provides practical tips for Oracle Business Intelligence Applications 11g implementations. It discusses scripting installations and configurations, LDAP integration challenges, implementing high availability, different methods for data extracts, and simplifying disaster recovery. Specific tips include scripting all processes, configuring the ODI agent JVM and connection pools for performance, understanding external LDAP authentication in ODI, implementing active-active high availability for ODI agents, choosing the right data extract method based on latency and volume, and using DataGuard and CNAMEs to simplify failover for disaster recovery.
OBIEE12c and Embedded Essbase 12c - An Initial Look at Query Acceleration Use...Mark Rittman
OBIEE12c comes with an updated version of Essbase that focuses entirely in this release on the query acceleration use-case. This presentation looks at this new release and explains how the new BI Accelerator Wizard manages the creation of Essbase cubes to accelerate OBIEE query performance
Tame Big Data with Oracle Data IntegrationMichael Rainey
In this session, Oracle Product Management covers how Oracle Data Integrator and Oracle GoldenGate are vital to big data initiatives across the enterprise, providing the movement, translation, and transformation of information and data not only heterogeneously but also in big data environments. Through a metadata-focused approach for cataloging, defining, and reusing big data technologies such as Hive, Hadoop Distributed File System (HDFS), HBase, Sqoop, Pig, Oracle Loader for Hadoop, Oracle SQL Connector for Hadoop Distributed File System, and additional big data projects, Oracle Data Integrator bridges the gap in the ability to unify data across these systems and helps deliver timely and trusted data to analytic and decision support platforms.
Co-presented with Alex Kotopoulis at Oracle OpenWorld 2014.
Reaching scale limits on a Hadoop platform: issues and errors created by spee...DataWorks Summit
Santander UK’s Big Data journey began in 2014, using Hadoop to make the most of our data and generate value for customers. Within 9 months, we created a highly available real-time customer facing application for customer analytics. We currently have 500 different people doing their own analysis and projects with this data, spanning a total of 50 different use cases. This data, (consisting of over 40 million customer records with billions of transactions), provides our business new insights that were inaccessible before.
Our business moves quickly, with several products and 20 use cases currently in production. We currently have a customer data lake and a technical data lake. Having a platform with very different workloads has proven to be challenging.
Our success in generating value created such growth in terms of data, use cases, analysts and usage patterns that 3 years later we find issues with scalability in HDFS, Hive metastore and Hadoop operations and challenges with highly available architectures with Hbase, Flume and Kafka. Going forward we are exploring alternative architectures including a hybrid cloud model, and moving towards streaming.
Our goal with this session is to assist people in the early part of their journey by building a solid foundation. We hope that others can benefit from us sharing our experiences and lessons learned during our journey.
Speaker
Nicolette Bullivant, Head of Data Engineering at Santander UK Technology, Santander UK Technology
New World Hadoop Architectures (& What Problems They Really Solve) for Oracle...Rittman Analytics
Most DBAs are aware something interesting is going on with big data and the Hadoop product ecosystem that underpins it, but aren't so clear about what each component in the stack does, what problem each part solves and why those problems couldn't be solved using the old approach. We'll look at where it's all going with the advent of Spark and machine learning, what's happening with ETL, metadata and analytics on this platform ... why IaaS and datawarehousing-as-a-service will have such a big impact, sooner than you think
This document discusses strategies for successfully utilizing a data lake. It notes that creating a data lake is just the beginning and that challenges include data governance, metadata management, access, and effective use of the data. The document advocates for data democratization through discovery, accessibility, and usability. It also discusses best practices like self-service BI and automated workload migration from data warehouses to reduce costs and risks. The key is to address the "data lake dilemma" of these challenges to avoid a "data swamp" and slow adoption.
The document provides an overview of Apache Atlas, a metadata management and governance solution for Hadoop data lakes. It discusses Atlas' architecture, which uses a graph database to store types and instances. Atlas also includes search capabilities and integration with Hadoop components like Hive to capture lineage metadata. The remainder of the document outlines Atlas' roadmap, with goals of adding additional component connectors, a governance certification program, and generally moving towards a production release.
This document discusses leveraging Hadoop within the existing data warehouse environment of the Department of Immigration and Border Protection (DIBP) in Australia. It provides an overview of DIBP's business and why Hadoop was adopted, describes the existing EDW environment, and discusses the technical implementation of Hadoop. It also outlines next steps such as consolidating the departmental EDW and advanced analytics on Hadoop, and concludes by taking questions.
GoldenGate and ODI - A Perfect Match for Real-Time Data WarehousingMichael Rainey
Oracle Data Integrator and Oracle GoldenGate excel as standalone products, but paired together they are the perfect match for real-time data warehousing. Following Oracle’s Next Generation Reference Data Warehouse Architecture, this discussion will provide best practices on how to configure, implement, and process data in real-time using ODI and GoldenGate. Attendees will see common real-time challenges solved, including parent-child relationships within micro-batch ETL.
Presented at RMOUG Training Days 2013 & KScope13.
This document discusses architecting Hadoop for adoption and data applications. It begins by explaining how traditional systems struggle as data volumes increase and how Hadoop can help address this issue. Potential Hadoop use cases are presented such as file archiving, data analytics, and ETL offloading. Total cost of ownership (TCO) is discussed for each use case. The document then covers important considerations for deploying Hadoop such as hardware selection, team structure, and impact across the organization. Lastly, it discusses lessons learned and the need for self-service tools going forward.
This document provides an overview of big data concepts and technologies for managers. It discusses problems with relational databases for large, unstructured data and introduces NoSQL databases and Hadoop as solutions. It also summarizes common big data applications, frameworks like MapReduce, Spark, and Flink, and different NoSQL database categories including key-value, column-family, document, and graph stores.
GoldenGate and Oracle Data Integrator - A Perfect Match- Upgrade to 12cMichael Rainey
- The document discusses upgrading Oracle GoldenGate 11g and Oracle Data Integrator 11g to their 12c versions. It provides an overview of the steps to upgrade each product including preparing the source and target systems, installing 12c, updating supplemental logging, and finalizing the upgrade by altering processes to write a new sequence number. It also discusses using the convprm tool to convert GoldenGate parameter files during the upgrade process.
Data Integration for Big Data (OOW 2016, Co-Presented With Oracle)Rittman Analytics
Oracle Data Integration Platform is a cornerstone for big data solutions that provides five core capabilities: business continuity, data movement, data transformation, data governance, and streaming data handling. It includes eight core products that can operate in the cloud or on-premise, and is considered the most innovative in areas like real-time/streaming integration and extract-load-transform capabilities with big data technologies. The platform offers a comprehensive architecture covering key areas like data ingestion, preparation, streaming integration, parallel connectivity, and governance.
The Rise of Big Data Governance: Insight on this Emerging Trend from Active O...DataWorks Summit
Each of today’s most forward-thinking enterprises have been forced to face similar data challenges: the reliance on real-time data to better serve their customers and, subsequently, the requirement of complying with regulations to protect that data – one example being the General Data Protection Regulation (GDPR).
The solution to this emerging challenge is a tricky one – for companies like ING, this data governance challenge has been met with metadata, a consistent view across a large heterogeneous ecosystem and collaboration with an active open source community.
This joint presentation, John Mertic – director of program management for ODPi – and Ferd Scheepers – Global Chief Information Architect of ING – will address the benefits of a vendor-neutral approach to data governance, the need for an open metadata standard, along with insight around how companies ING, IBM, Hortonworks and more are delivering solutions to this challenge as an open source initiative.
Speakers
John Mertic, Director of Program Management for ODPi, R Consortium, and Open Mainframe Project, The Linux Foundation
Maryna Strelchuk, Information Architect, ING
Oracle PL/SQL 12c and 18c New Features + RADstack + Community SitesSteven Feuerstein
Slides presented at moug.org's August 2018 conference. Covers the RADstack (REST - APEX - Database) + our community sites (AskTOM, LiveSQL and Dev Gym) + a whole bunch of cool new PL/SQL features. Search LiveSQL.oracle.com for scripts to match up with the features presented.
451 Research is a leading IT research and advisory company founded in 2000 with over 250 employees including over 100 analysts. It provides research and data through fifteen channels to over 1,000 clients on technology and service providers. The document discusses the evolution of the meaning of "Hadoop" from referring originally to specific Apache projects like HDFS and MapReduce to becoming a catch-all term for the distributed data processing ecosystem, and how different Hadoop distributions combine various related Apache projects in their offerings. It also examines how data platforms are converging, with various databases, analytics engines, and streaming platforms increasingly supporting common workloads and data models.
Obiee 12C and the Leap Forward in Lifecycle ManagementStewart Bryson
One reason that OBIEE projects regularly languish between design and delivery is the product's historical lack of concrete lifecycle and deployment capabilities. Whether adopting manual lifecycle processes, or scripting complex deployments, these solutions usually yield mixed results, but always highlight one clear conclusion: we need change.
With the newest release of OBIEE 12c, change has finally arrived. This session will detail the new BI Application Archive (BAR) file and it's role in smoother migrations and increased developer productivity. We'll take a peak under the covers at a new flexible, container-like architecture that makes the creation and application of these BAR files simple, but also provides enhanced lifecycle capabilities throughout.
Effective data governance is imperative to the success of Data Lake initiatives. Without governance policies and processes, information discovery and analysis is severely impaired. In this session we will provide an in-depth look into the Data Governance Initiative launched collaboratively between Hortonworks and partners from across industries. We will cover the objectives of Data Governance Initiatives and demonstrate key governance capabilities of the Hortonworks Data Platform.
Accelerating query processing with materialized views in Apache HiveDataWorks Summit
Over the last few years, the Apache Hive community has been working on advancements to enable a full new range of use cases for the project, moving from its batch processing roots towards a SQL interactive query answering platform. Traditionally, one of the most powerful techniques used to accelerate query processing in data warehouses is the precomputation of relevant summaries or materialized views.
This talk presents our work on introducing materialized views and automatic query rewriting based on those materializations in Apache Hive. In particular, materialized views can be stored natively in Hive or in other systems such as Druid using custom storage handlers, and they can seamlessly exploit new exciting Hive features such as LLAP acceleration. Then the optimizer relies in Apache Calcite to automatically produce full and partial rewritings for a large set of query expressions comprising projections, filters, join, and aggregation operations. We shall describe the current coverage of the rewriting algorithm, how Hive controls important aspects of the life cycle of the materialized views such as the freshness of their data, and outline interesting directions for future improvements. We include an experimental evaluation highlighting the benefits that the usage of materialized views can bring to the execution of Hive workloads.
Speaker
Jesus Camacho Rodriguez, Member of Technical Staff, Hortonworks
The world’s largest enterprises run their infrastructure on Oracle, DB2 and SQL and their critical business operations on SAP applications. Organisations need this data to be available in real-time to conduct necessary analytics. However, delivering this heterogeneous data at the speed it’s required can be a huge challenge because of the complex underlying data models and structures and legacy manual processes which are prone to errors and delays.
Unlock these silos of data and enable the new advanced analytics platforms by attending this session.
Find out how to:
• To overcome common challenges faced by enterprises trying to access their SAP data
• You can integrate SAP data in real-time with change data capture (CDC) technology
• Organisations are using Attunity Replicate for SAP to stream SAP data in to Kafka
My presentation slides from Hadoop Summit, San Jose, June 28, 2016. See live video at http://www.makedatauseful.com/vid-solving-performance-problems-hadoop/ and follow along for context.
Moving analytic workloads into production - specific technical challenges and best practices for engineering SQL in Hadoop solutions. Highlighting the next generation engineering approaches to the secret sauce we have implemented in the Actian VectorH database.
The document discusses Oracle's big data platform and how it can extend Hortonworks' data platform. It provides an overview of Oracle's enterprise big data architecture and the key components of its big data platform. It also discusses how Oracle's platform provides rich SQL access across different data sources and describes some big data solutions for adaptive marketing and predictive maintenance.
The document discusses extending data governance in Hadoop ecosystems using Apache Atlas and partner solutions including Waterline Data, Attivo, and Trifacta. It highlights how these vendors have adopted Apache's open source community commitment and are integrating their products with Atlas to provide a rich, innovative community with a common metadata store backed by Atlas. The session will showcase how these three vendors extend governance capabilities by integrating their products with Atlas.
This document discusses Oracle Data Integration solutions for tapping into big data reservoirs. It begins with an overview of Oracle Data Integration and how it can improve agility, reduce risk and costs. It then discusses Oracle's approach to comprehensive data integration and governance capabilities including real-time data movement, data transformation, data federation, and more. The document also provides examples of how Oracle Data Integration has been used by customers for big data use cases involving petabytes of data.
The document discusses Oracle's data integration products and big data solutions. It outlines five core capabilities of Oracle's data integration platform, including data availability, data movement, data transformation, data governance, and streaming data. It then describes eight core products that address real-time and streaming integration, ELT integration, data preparation, streaming analytics, dataflow ML, metadata management, data quality, and more. The document also outlines five cloud solutions for data integration including data migrations, data warehouse integration, development and test environments, high availability, and heterogeneous cloud. Finally, it discusses pragmatic big data solutions for data ingestion, transformations, governance, connectors, and streaming big data.
The document discusses new features and enhancements in Oracle Business Intelligence 12c, including:
- Improved front-end usability features like an updated home page, enhanced sorting, and view properties access in compound layouts.
- New data visualization options like Oracle Data Visualization for ad-hoc analysis and mashups of spreadsheets with existing subject areas.
- Changes to installation, configuration, and architecture like a simplified installation process and separation of environment metadata from configuration.
- Upgrades are supported from 11g using a new Baseline Validation Tool to test and compare systems before and after upgrades.
The Time is Now! Migrating from OWB to ODI 12cStewart Bryson
Prior to the introduction of Data Integrator (ODI), Oracle had another data integration tool: Warehouse Builder (OWB). Usually positioned as an ETL tool, OWB excelled in environments with a strong footprint in the Oracle Database. Oracle's statement of direction has been clear: to deliver a unified data integration platform, combining the best from both tools into a true world class product. With ODI 12c, that day has arrived.
In this presentation, I’ll demonstrate the features available for migrating from OWB to ODI 12c. I’ll also describe a phased approach for doing a “right-time” conversion to ODI 12c, which involves migrating bite-sized chunks of OWB processes over to ODI when that migration adds legitimate value for the customer.
This document discusses strategies for successfully utilizing a data lake. It notes that creating a data lake is just the beginning and that challenges include data governance, metadata management, access, and effective use of the data. The document advocates for data democratization through discovery, accessibility, and usability. It also discusses best practices like self-service BI and automated workload migration from data warehouses to reduce costs and risks. The key is to address the "data lake dilemma" of these challenges to avoid a "data swamp" and slow adoption.
The document provides an overview of Apache Atlas, a metadata management and governance solution for Hadoop data lakes. It discusses Atlas' architecture, which uses a graph database to store types and instances. Atlas also includes search capabilities and integration with Hadoop components like Hive to capture lineage metadata. The remainder of the document outlines Atlas' roadmap, with goals of adding additional component connectors, a governance certification program, and generally moving towards a production release.
This document discusses leveraging Hadoop within the existing data warehouse environment of the Department of Immigration and Border Protection (DIBP) in Australia. It provides an overview of DIBP's business and why Hadoop was adopted, describes the existing EDW environment, and discusses the technical implementation of Hadoop. It also outlines next steps such as consolidating the departmental EDW and advanced analytics on Hadoop, and concludes by taking questions.
GoldenGate and ODI - A Perfect Match for Real-Time Data WarehousingMichael Rainey
Oracle Data Integrator and Oracle GoldenGate excel as standalone products, but paired together they are the perfect match for real-time data warehousing. Following Oracle’s Next Generation Reference Data Warehouse Architecture, this discussion will provide best practices on how to configure, implement, and process data in real-time using ODI and GoldenGate. Attendees will see common real-time challenges solved, including parent-child relationships within micro-batch ETL.
Presented at RMOUG Training Days 2013 & KScope13.
This document discusses architecting Hadoop for adoption and data applications. It begins by explaining how traditional systems struggle as data volumes increase and how Hadoop can help address this issue. Potential Hadoop use cases are presented such as file archiving, data analytics, and ETL offloading. Total cost of ownership (TCO) is discussed for each use case. The document then covers important considerations for deploying Hadoop such as hardware selection, team structure, and impact across the organization. Lastly, it discusses lessons learned and the need for self-service tools going forward.
This document provides an overview of big data concepts and technologies for managers. It discusses problems with relational databases for large, unstructured data and introduces NoSQL databases and Hadoop as solutions. It also summarizes common big data applications, frameworks like MapReduce, Spark, and Flink, and different NoSQL database categories including key-value, column-family, document, and graph stores.
GoldenGate and Oracle Data Integrator - A Perfect Match- Upgrade to 12cMichael Rainey
- The document discusses upgrading Oracle GoldenGate 11g and Oracle Data Integrator 11g to their 12c versions. It provides an overview of the steps to upgrade each product including preparing the source and target systems, installing 12c, updating supplemental logging, and finalizing the upgrade by altering processes to write a new sequence number. It also discusses using the convprm tool to convert GoldenGate parameter files during the upgrade process.
Data Integration for Big Data (OOW 2016, Co-Presented With Oracle)Rittman Analytics
Oracle Data Integration Platform is a cornerstone for big data solutions that provides five core capabilities: business continuity, data movement, data transformation, data governance, and streaming data handling. It includes eight core products that can operate in the cloud or on-premise, and is considered the most innovative in areas like real-time/streaming integration and extract-load-transform capabilities with big data technologies. The platform offers a comprehensive architecture covering key areas like data ingestion, preparation, streaming integration, parallel connectivity, and governance.
The Rise of Big Data Governance: Insight on this Emerging Trend from Active O...DataWorks Summit
Each of today’s most forward-thinking enterprises have been forced to face similar data challenges: the reliance on real-time data to better serve their customers and, subsequently, the requirement of complying with regulations to protect that data – one example being the General Data Protection Regulation (GDPR).
The solution to this emerging challenge is a tricky one – for companies like ING, this data governance challenge has been met with metadata, a consistent view across a large heterogeneous ecosystem and collaboration with an active open source community.
This joint presentation, John Mertic – director of program management for ODPi – and Ferd Scheepers – Global Chief Information Architect of ING – will address the benefits of a vendor-neutral approach to data governance, the need for an open metadata standard, along with insight around how companies ING, IBM, Hortonworks and more are delivering solutions to this challenge as an open source initiative.
Speakers
John Mertic, Director of Program Management for ODPi, R Consortium, and Open Mainframe Project, The Linux Foundation
Maryna Strelchuk, Information Architect, ING
Oracle PL/SQL 12c and 18c New Features + RADstack + Community SitesSteven Feuerstein
Slides presented at moug.org's August 2018 conference. Covers the RADstack (REST - APEX - Database) + our community sites (AskTOM, LiveSQL and Dev Gym) + a whole bunch of cool new PL/SQL features. Search LiveSQL.oracle.com for scripts to match up with the features presented.
451 Research is a leading IT research and advisory company founded in 2000 with over 250 employees including over 100 analysts. It provides research and data through fifteen channels to over 1,000 clients on technology and service providers. The document discusses the evolution of the meaning of "Hadoop" from referring originally to specific Apache projects like HDFS and MapReduce to becoming a catch-all term for the distributed data processing ecosystem, and how different Hadoop distributions combine various related Apache projects in their offerings. It also examines how data platforms are converging, with various databases, analytics engines, and streaming platforms increasingly supporting common workloads and data models.
Obiee 12C and the Leap Forward in Lifecycle ManagementStewart Bryson
One reason that OBIEE projects regularly languish between design and delivery is the product's historical lack of concrete lifecycle and deployment capabilities. Whether adopting manual lifecycle processes, or scripting complex deployments, these solutions usually yield mixed results, but always highlight one clear conclusion: we need change.
With the newest release of OBIEE 12c, change has finally arrived. This session will detail the new BI Application Archive (BAR) file and it's role in smoother migrations and increased developer productivity. We'll take a peak under the covers at a new flexible, container-like architecture that makes the creation and application of these BAR files simple, but also provides enhanced lifecycle capabilities throughout.
Effective data governance is imperative to the success of Data Lake initiatives. Without governance policies and processes, information discovery and analysis is severely impaired. In this session we will provide an in-depth look into the Data Governance Initiative launched collaboratively between Hortonworks and partners from across industries. We will cover the objectives of Data Governance Initiatives and demonstrate key governance capabilities of the Hortonworks Data Platform.
Accelerating query processing with materialized views in Apache HiveDataWorks Summit
Over the last few years, the Apache Hive community has been working on advancements to enable a full new range of use cases for the project, moving from its batch processing roots towards a SQL interactive query answering platform. Traditionally, one of the most powerful techniques used to accelerate query processing in data warehouses is the precomputation of relevant summaries or materialized views.
This talk presents our work on introducing materialized views and automatic query rewriting based on those materializations in Apache Hive. In particular, materialized views can be stored natively in Hive or in other systems such as Druid using custom storage handlers, and they can seamlessly exploit new exciting Hive features such as LLAP acceleration. Then the optimizer relies in Apache Calcite to automatically produce full and partial rewritings for a large set of query expressions comprising projections, filters, join, and aggregation operations. We shall describe the current coverage of the rewriting algorithm, how Hive controls important aspects of the life cycle of the materialized views such as the freshness of their data, and outline interesting directions for future improvements. We include an experimental evaluation highlighting the benefits that the usage of materialized views can bring to the execution of Hive workloads.
Speaker
Jesus Camacho Rodriguez, Member of Technical Staff, Hortonworks
The world’s largest enterprises run their infrastructure on Oracle, DB2 and SQL and their critical business operations on SAP applications. Organisations need this data to be available in real-time to conduct necessary analytics. However, delivering this heterogeneous data at the speed it’s required can be a huge challenge because of the complex underlying data models and structures and legacy manual processes which are prone to errors and delays.
Unlock these silos of data and enable the new advanced analytics platforms by attending this session.
Find out how to:
• To overcome common challenges faced by enterprises trying to access their SAP data
• You can integrate SAP data in real-time with change data capture (CDC) technology
• Organisations are using Attunity Replicate for SAP to stream SAP data in to Kafka
My presentation slides from Hadoop Summit, San Jose, June 28, 2016. See live video at http://www.makedatauseful.com/vid-solving-performance-problems-hadoop/ and follow along for context.
Moving analytic workloads into production - specific technical challenges and best practices for engineering SQL in Hadoop solutions. Highlighting the next generation engineering approaches to the secret sauce we have implemented in the Actian VectorH database.
The document discusses Oracle's big data platform and how it can extend Hortonworks' data platform. It provides an overview of Oracle's enterprise big data architecture and the key components of its big data platform. It also discusses how Oracle's platform provides rich SQL access across different data sources and describes some big data solutions for adaptive marketing and predictive maintenance.
The document discusses extending data governance in Hadoop ecosystems using Apache Atlas and partner solutions including Waterline Data, Attivo, and Trifacta. It highlights how these vendors have adopted Apache's open source community commitment and are integrating their products with Atlas to provide a rich, innovative community with a common metadata store backed by Atlas. The session will showcase how these three vendors extend governance capabilities by integrating their products with Atlas.
This document discusses Oracle Data Integration solutions for tapping into big data reservoirs. It begins with an overview of Oracle Data Integration and how it can improve agility, reduce risk and costs. It then discusses Oracle's approach to comprehensive data integration and governance capabilities including real-time data movement, data transformation, data federation, and more. The document also provides examples of how Oracle Data Integration has been used by customers for big data use cases involving petabytes of data.
The document discusses Oracle's data integration products and big data solutions. It outlines five core capabilities of Oracle's data integration platform, including data availability, data movement, data transformation, data governance, and streaming data. It then describes eight core products that address real-time and streaming integration, ELT integration, data preparation, streaming analytics, dataflow ML, metadata management, data quality, and more. The document also outlines five cloud solutions for data integration including data migrations, data warehouse integration, development and test environments, high availability, and heterogeneous cloud. Finally, it discusses pragmatic big data solutions for data ingestion, transformations, governance, connectors, and streaming big data.
The document discusses new features and enhancements in Oracle Business Intelligence 12c, including:
- Improved front-end usability features like an updated home page, enhanced sorting, and view properties access in compound layouts.
- New data visualization options like Oracle Data Visualization for ad-hoc analysis and mashups of spreadsheets with existing subject areas.
- Changes to installation, configuration, and architecture like a simplified installation process and separation of environment metadata from configuration.
- Upgrades are supported from 11g using a new Baseline Validation Tool to test and compare systems before and after upgrades.
The Time is Now! Migrating from OWB to ODI 12cStewart Bryson
Prior to the introduction of Data Integrator (ODI), Oracle had another data integration tool: Warehouse Builder (OWB). Usually positioned as an ETL tool, OWB excelled in environments with a strong footprint in the Oracle Database. Oracle's statement of direction has been clear: to deliver a unified data integration platform, combining the best from both tools into a true world class product. With ODI 12c, that day has arrived.
In this presentation, I’ll demonstrate the features available for migrating from OWB to ODI 12c. I’ll also describe a phased approach for doing a “right-time” conversion to ODI 12c, which involves migrating bite-sized chunks of OWB processes over to ODI when that migration adds legitimate value for the customer.
AgileAnalytics: Agile Real-Time BI with Oracle Business Intelligence, Oracle ...Stewart Bryson
A successful real-time BI implementation involves integrating the logical and physical model, source and target databases, ETL subsystems, and front-end dashboards and reports. Agile methodologies often struggle to deliver real-time implementations due to the complexity of the intermingled moving parts.
Using the Oracle Information Management Reference Architecture as a backdrop, attendees will see how to construct a complete real-time BI system using Oracle Business Intelligence, Oracle Data Integrator and Oracle GoldenGate, driven by an Agile approach focusing on user stories and time-boxed iterations.
The Time is Now: Migrating from Oracle Warehouse Builder to Oracle Data Integ...Stewart Bryson
Prior to the introduction of Data Integrator (ODI), Oracle had another data integration tool: Warehouse Builder (OWB). Usually positioned as an ETL tool, OWB excelled in environments with a strong footprint in the Oracle Database. Oracle's statement of direction has been clear: to deliver a unified data integration platform, combining the best from both tools into a true world class product. With ODI 12c, that day has arrived.
In this presentation, I’ll demonstrate the features available for migrating from OWB to ODI 12c. I’ll also describe a phased approach for doing a “right-time” conversion to ODI 12c, which involves migrating bite-sized chunks of OWB processes over to ODI when that migration adds legitimate value for the customer.
An Introduction to Oracle Enterprise Metadata ManagerLauren Prezby
This document discusses Red Pill Analytics and its Oracle Enterprise Metadata Manager (OEMM) product. It provides an overview of OEMM's key capabilities including harvesting metadata from various sources to create models, using bridges to connect the models, and stitching the models together to form relationships and lineage. Use cases for OEMM are also mentioned such as supporting agile delivery, upgrading Oracle BI, understanding big data assets, and aiding project planning.
Obiee 12c and the leap forward in lifecycle managementLauren Prezby
This document discusses Oracle's OBIEE 12c and improvements to lifecycle management. It proposes a more flexible architecture using pluggable metadata modules and lightweight service instances. This would allow customizing metadata independently of other instances and simplify content portability using BAR files. However, the current 12c release only supports a singleton instance configuration.
This document summarizes a presentation about using Hadoop as an analytic platform. It discusses how Actian has added seven key ingredients to Hadoop to unlock its full potential for analytics. These include high-speed data integration, a visual framework for data science and modeling, open-source analytic operators, high-performance data processing engines, vector-based SQL processing natively on HDFS, an extremely fast parallel analytics engine, and a next-generation big data analytics platform. The goal is to transform Hadoop from merely a data reservoir to a fully-featured analytics platform.
Can you Re-Platform your Teradata, Oracle, Netezza and SQL Server Analytic Wo...DataWorks Summit
The document discusses re-platforming existing enterprise business intelligence and analytic workloads from platforms like Oracle, Teradata, SAP and IBM to the Hadoop platform. It notes that many existing analytic workloads are struggling with increasing data volumes and are too costly. Hadoop offers a modern distributed platform that can address these issues through the use of a production-grade SQL database like VectorH on Hadoop. The document provides guidelines for re-platforming workloads and notes potential benefits such as improved performance, reduced costs and leveraging the Hadoop ecosystem.
This presentation is for Analytic and Business Intelligence leads as well as IT leads who manage analytics. In addition, existing Oracle Business Intelligence and Analytic Customers will find it valuable to understand how they can leverage their existing investments along with Oracle Analytics Cloud.
BI congres 2016-2: Diving into weblog data with SAS on Hadoop - Lisa Truyers...BICC Thomas More
9de BI congres van het BICC-Thomas More: 24 maart 2016
De hoeveelheid data die via weblogs verzameld wordt, neemt steeds meer toe. Lisa Truyers zet aan de hand van een praktische case uiteen hoe Keyrus hiermee aan de slag ging
The document discusses an upcoming webinar on Big Data and SQL. It provides details on the webinar topics, speakers, and the HP Vertica Analytics Platform. The webinar will explore how HP Vertica allows users to navigate and analyze data stored in Hadoop using SQL, avoiding complex ETL processes. It will also discuss how the platform handles both query and analytical workloads and enables exploration of semi-structured data through its Flex Zone.
Pivotal Big Data Suite: A Technical OverviewVMware Tanzu
Pivotal provides a suite of big data products including Pivotal Greenplum Database, Pivotal HDB, and Pivotal GemFire. Greenplum Database is an open source massively parallel processing data warehouse. HDB is an open source analytical database for Apache Hadoop. GemFire is an open source application and transaction data grid. The suite provides a complete platform for big data with deployment options, advanced data services, and flexible licensing.
The document discusses how MySQL can be used to unlock insights from big data. It describes how MySQL provides both SQL and NoSQL access to data stored in Hadoop, allowing organizations to analyze large, diverse datasets. Tools like Apache Sqoop and the MySQL Applier for Hadoop are used to import data from MySQL to Hadoop for advanced analytics, while solutions like MySQL Fabric allow databases to scale out through data sharding.
Oracle Openworld Presentation with Paul Kent (SAS) on Big Data Appliance and ...jdijcks
Learn about the benefits of Oracle Big Data Appliance and how it can drive business value underneath applications and tools. This includes a section by Paul Kent, VP Big Data SAS describing how SAS runs well on Oracle Engineered Systems and on Oracle Big Data Appliance specifically.
Oracle Unified Information Architeture + Analytics by ExampleHarald Erb
Der Vortrag gibt zunächst einen Architektur-Überblick zu den UIA-Komponenten und deren Zusammenspiel. Anhand eines Use Cases wird vorgestellt, wie im "UIA Data Reservoir" einerseits kostengünstig aktuelle Daten "as is" in einem Hadoop File System (HDFS) und andererseits veredelte Daten in einem Oracle 12c Data Warehouse miteinander kombiniert oder auch per Direktzugriff in Oracle Business Intelligence ausgewertet bzw. mit Endeca Information Discovery auf neue Zusammenhänge untersucht werden.
Big Data in Action – Real-World Solution ShowcaseInside Analysis
The Briefing Room with Radiant Advisors and IBM
Live Webcast on February 25, 2014
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=53c9b7fa2000f98f5b236747e3602511
The power of Big Data depends heavily upon the context in which it's used, and most organizations are just beginning to figure out where, how and when to leverage it. One key to success is integration with existing information systems, many of which still rely on relational database technologies. Finding ways to blend these two worlds can help companies generate measurable business value in fairly short order.
Register for this episode of The Briefing Room to hear Analysts Lindy Ryan and John O'Brien as they explain how the combination of traditional Business Intelligence with Big Data Analytics can provide game-changing results in today's information economy. They'll be briefed by Eric Poulin and Paul Flach of Stream Integration who will share best practices for designing and implementing Big Data solutions. They'll discuss the components of IBM BigInsights, and explain how BigSheets can empower non-technical users who need to explore self-structured data.
Visit InsideAnlaysis.com for more information.
Game Changed – How Hadoop is Reinventing Enterprise ThinkingInside Analysis
The Briefing Room with Dr. Robin Bloor and RedPoint Global
Live Webcast on April 8, 2014
Watch the archive: https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=cfa1bffdd62dc6677fa225bdffe4a0b9
The innovation curve often arcs slowly before picking up speed. Companies that harness a major transformation early in the game can make serious headway before challengers enter the picture. The world of Hadoop features several of these upstarts, each of which uses the open-source foundation as an engine to drive vastly greater performance to a wide range of services, and even create new ones.
Register for this episode of The Briefing Room to hear veteran Analyst Dr. Robin Bloor explain how the Hadoop engine is being used to architect a new generation of enterprise applications. He’ll be briefed by George Corugedo, RedPoint Global CTO and Co-founder, who will showcase how enterprises can cost-effectively take advantage of the scalability, processing power and lower costs that Hadoop 2.0/YARN applications offer by eliminating the long-term expense of hiring MapReduce programmers.
Visit InsideAnlaysis.com for more information.
Level Up – How to Achieve Hadoop AccelerationInside Analysis
The Briefing Room with Robin Bloor and HP Vertica
Live Webcast on August 26, 2014
Watch the archive:
https://bloorgroup.webex.com/bloorgroup/lsr.php?RCID=3dd6d1b068fe395f665c75adb682ac41
Hadoop has long passed the point of being a nascent technology, but many users have found that when left to its own devices, Hadoop can be a one trick pony. To get the most out of Hadoop, organizations need a flexible platform that empowers analysts and data managers with a complete set of information lifecycle management and analytics tools without a performance tradeoff.
Register for this episode of The Briefing Room to hear veteran Analyst Dr. Robin Bloor as he outlines Hadoop’s role in a big data architecture. He’ll be briefed by Walt Maguire of HP Vertica, who will showcase his company’s big data solutions, including HAVEn and the HP Big Data Platform. He will demonstrate how HP Vertica acts as a complement to Hadoop, and how the combination of the two provides a versatile and highly performant solution.
Visit InsideAnlaysis.com for more information.
Hadoop for Business Intelligence ProfessionalsSkillspeed
This is a presentation on Hadoop for BI Professionals who want to upgrade their career path to BIG Data technologies. Hadoop for Business Intelligence Professionals is a definite upgrade in terms of career growth, scope of worth and organization influence.
The PPT covers the following topics:
✓ What is BIG Data?
✓ What is Hadoop? Why is it so popular?
✓ Upgrading from BI to Hadoop
✓ Career Path
✓ Salary & Job Trends
✓ Hiring Companies
----------
Skillspeed is a live e-learning company focusing on high-technology courses. We provide live instructor led training in BIG Data & Hadoop featuring Realtime Projects, 24/7 Lifetime Support & 100% Placement Assistance.
Email: [email protected]
Website: https://www.skillspeed.com
A modern, flexible approach to Hadoop implementation incorporating innovation...DataWorks Summit
A modern, flexible approach to Hadoop implementation incorporating innovations from HP Haven
Jeff Veis
Vice President
HP Software Big Data
Gilles Noisette
Master Solution Architect
HP EMEA Big Data CoE
Automation Dreamin' 2022: Sharing Some Gratitude with Your UsersLynda Kane
Slide Deck from Automation Dreamin'2022 presentation Sharing Some Gratitude with Your Users on creating a Flow to present a random statement of Gratitude to a User in Salesforce.
Mobile App Development Company in Saudi ArabiaSteve Jonas
EmizenTech is a globally recognized software development company, proudly serving businesses since 2013. With over 11+ years of industry experience and a team of 200+ skilled professionals, we have successfully delivered 1200+ projects across various sectors. As a leading Mobile App Development Company In Saudi Arabia we offer end-to-end solutions for iOS, Android, and cross-platform applications. Our apps are known for their user-friendly interfaces, scalability, high performance, and strong security features. We tailor each mobile application to meet the unique needs of different industries, ensuring a seamless user experience. EmizenTech is committed to turning your vision into a powerful digital product that drives growth, innovation, and long-term success in the competitive mobile landscape of Saudi Arabia.
Leading AI Innovation As A Product Manager - Michael JidaelMichael Jidael
Unlike traditional product management, AI product leadership requires new mental models, collaborative approaches, and new measurement frameworks. This presentation breaks down how Product Managers can successfully lead AI Innovation in today's rapidly evolving technology landscape. Drawing from practical experience and industry best practices, I shared frameworks, approaches, and mindset shifts essential for product leaders navigating the unique challenges of AI product development.
In this deck, you'll discover:
- What AI leadership means for product managers
- The fundamental paradigm shift required for AI product development.
- A framework for identifying high-value AI opportunities for your products.
- How to transition from user stories to AI learning loops and hypothesis-driven development.
- The essential AI product management framework for defining, developing, and deploying intelligence.
- Technical and business metrics that matter in AI product development.
- Strategies for effective collaboration with data science and engineering teams.
- Framework for handling AI's probabilistic nature and setting stakeholder expectations.
- A real-world case study demonstrating these principles in action.
- Practical next steps to begin your AI product leadership journey.
This presentation is essential for Product Managers, aspiring PMs, product leaders, innovators, and anyone interested in understanding how to successfully build and manage AI-powered products from idea to impact. The key takeaway is that leading AI products is about creating capabilities (intelligence) that continuously improve and deliver increasing value over time.
AI EngineHost Review: Revolutionary USA Datacenter-Based Hosting with NVIDIA ...SOFTTECHHUB
I started my online journey with several hosting services before stumbling upon Ai EngineHost. At first, the idea of paying one fee and getting lifetime access seemed too good to pass up. The platform is built on reliable US-based servers, ensuring your projects run at high speeds and remain safe. Let me take you step by step through its benefits and features as I explain why this hosting solution is a perfect fit for digital entrepreneurs.
Procurement Insights Cost To Value Guide.pptxJon Hansen
Procurement Insights integrated Historic Procurement Industry Archives, serves as a powerful complement — not a competitor — to other procurement industry firms. It fills critical gaps in depth, agility, and contextual insight that most traditional analyst and association models overlook.
Learn more about this value- driven proprietary service offering here.
AI Changes Everything – Talk at Cardiff Metropolitan University, 29th April 2...Alan Dix
Talk at the final event of Data Fusion Dynamics: A Collaborative UK-Saudi Initiative in Cybersecurity and Artificial Intelligence funded by the British Council UK-Saudi Challenge Fund 2024, Cardiff Metropolitan University, 29th April 2025
https://alandix.com/academic/talks/CMet2025-AI-Changes-Everything/
Is AI just another technology, or does it fundamentally change the way we live and think?
Every technology has a direct impact with micro-ethical consequences, some good, some bad. However more profound are the ways in which some technologies reshape the very fabric of society with macro-ethical impacts. The invention of the stirrup revolutionised mounted combat, but as a side effect gave rise to the feudal system, which still shapes politics today. The internal combustion engine offers personal freedom and creates pollution, but has also transformed the nature of urban planning and international trade. When we look at AI the micro-ethical issues, such as bias, are most obvious, but the macro-ethical challenges may be greater.
At a micro-ethical level AI has the potential to deepen social, ethnic and gender bias, issues I have warned about since the early 1990s! It is also being used increasingly on the battlefield. However, it also offers amazing opportunities in health and educations, as the recent Nobel prizes for the developers of AlphaFold illustrate. More radically, the need to encode ethics acts as a mirror to surface essential ethical problems and conflicts.
At the macro-ethical level, by the early 2000s digital technology had already begun to undermine sovereignty (e.g. gambling), market economics (through network effects and emergent monopolies), and the very meaning of money. Modern AI is the child of big data, big computation and ultimately big business, intensifying the inherent tendency of digital technology to concentrate power. AI is already unravelling the fundamentals of the social, political and economic world around us, but this is a world that needs radical reimagining to overcome the global environmental and human challenges that confront us. Our challenge is whether to let the threads fall as they may, or to use them to weave a better future.
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptxJustin Reock
Building 10x Organizations with Modern Productivity Metrics
10x developers may be a myth, but 10x organizations are very real, as proven by the influential study performed in the 1980s, ‘The Coding War Games.’
Right now, here in early 2025, we seem to be experiencing YAPP (Yet Another Productivity Philosophy), and that philosophy is converging on developer experience. It seems that with every new method we invent for the delivery of products, whether physical or virtual, we reinvent productivity philosophies to go alongside them.
But which of these approaches actually work? DORA? SPACE? DevEx? What should we invest in and create urgency behind today, so that we don’t find ourselves having the same discussion again in a decade?
Learn the Basics of Agile Development: Your Step-by-Step GuideMarcel David
New to Agile? This step-by-step guide is your perfect starting point. "Learn the Basics of Agile Development" simplifies complex concepts, providing you with a clear understanding of how Agile can improve software development and project management. Discover the benefits of iterative work, team collaboration, and flexible planning.
TrustArc Webinar: Consumer Expectations vs Corporate Realities on Data Broker...TrustArc
Most consumers believe they’re making informed decisions about their personal data—adjusting privacy settings, blocking trackers, and opting out where they can. However, our new research reveals that while awareness is high, taking meaningful action is still lacking. On the corporate side, many organizations report strong policies for managing third-party data and consumer consent yet fall short when it comes to consistency, accountability and transparency.
This session will explore the research findings from TrustArc’s Privacy Pulse Survey, examining consumer attitudes toward personal data collection and practical suggestions for corporate practices around purchasing third-party data.
Attendees will learn:
- Consumer awareness around data brokers and what consumers are doing to limit data collection
- How businesses assess third-party vendors and their consent management operations
- Where business preparedness needs improvement
- What these trends mean for the future of privacy governance and public trust
This discussion is essential for privacy, risk, and compliance professionals who want to ground their strategies in current data and prepare for what’s next in the privacy landscape.
Hands On: Create a Lightning Aura Component with force:RecordDataLynda Kane
Slide Deck from the 3/26/2020 virtual meeting of the Cleveland Developer Group presentation on creating a Lightning Aura Component using force:RecordData.
#3: This is a Safe Harbor Front slide, one of two Safe Harbor Statement slides included in this template.
One of the Safe Harbor slides must be used if your presentation covers material affected by Oracle’s Revenue Recognition Policy
To learn more about this policy, e-mail: [email protected]