This document discusses real-time big data analytics from deployment to production. It covers:
1) Distilling raw data like log files and sensor streams into structured data using Hadoop for analytics.
2) Developing predictive models using techniques like decision trees, clustering, and ensembles on structured data.
3) Deploying models for real-time scoring via SQL, code, or PMML on either batch lookup tables or streaming data factors.
4) Scoring billions of predictions daily for applications like determining why customers buy products and attributing marketing channels.
5) Regularly refreshing models to incorporate new data and outcomes using techniques like exploratory analysis and time-to-event modeling
An API is an application programming interface that allows machines like apps and systems to access and use data and functionality over a network. APIs are intended for machine use rather than human use directly. Organizations use APIs to integrate their systems, reuse functionality, and increase automation. Individuals can use APIs to enhance their experiences and automate tasks. The API economy refers to the growing use of APIs, and API marketplaces help connect API providers and consumers.
This document provides an overview of big data analytics, strategies, and the WSO2 big data platform. It discusses how the amount of data in the world is growing exponentially due to factors like increased data collection and the internet of things. It then summarizes the WSO2 big data platform for collecting, processing, analyzing and visualizing large datasets. Key components include the complex event processor for query processing and the business activity monitor for dashboards. The document concludes by outlining new developments and features being worked on, such as distributed complex event processing and machine learning integration.
This video includes:
Purpose of Data Science, Role of Data Scientist, Skills required for Data Scientist, Job roles for Data Scientist, Applications of Data Science, Career in Data Science.
Want to pursue career in Data Science? Have knowledge of limited opportunities? Don't worry!
This e- book helps readers to know about top career opportunities one can pursue in Data Science. Further info.- https://www.henryharvin.com/business-analytics-course-with-python
This document discusses the potential of artificial intelligence (AI) and emerging technologies in healthcare. It begins with brief introductions and then outlines several key AI use cases in healthcare, including data collection and management, personal health data management, diagnosis, patient management, and macro health analysis. It also discusses challenges like skilled labor shortages and lack of large data sets. Risks of AI like bias, privacy issues, and dangerous mistakes are presented. The conclusion is that AI has great potential to transform healthcare if applications are handled carefully and data is managed appropriately.
This document discusses data science and big data. It begins by explaining how the volume of data from various sources has grown exponentially. It then defines data science as work dealing with collecting, preparing, analyzing, visualizing, managing and preserving large data collections. Big data is described as having four dimensions: volume, variety, velocity and veracity. Examples are given of how companies like Facebook and Google process huge amounts of data daily. The document discusses techniques like parallelization for dealing with big data volumes. Applications of big data are outlined across various industries. Programming languages and skills needed for data science are listed. Finally, the high career prospects and compensation for data scientists are highlighted.
The document discusses data science, defining it as a field that employs techniques from many areas like statistics, computer science, and mathematics to understand and analyze real-world phenomena. It explains that data science involves collecting, processing, and analyzing large amounts of data to discover patterns and make predictions. The document also notes that data science is an in-demand field that is expected to continue growing significantly in the coming years.
This was first part of the presentation on "Road Map for Careers in Big Data" in Conjunction with Hortonworks/Aengus Rooney on 17th August 2016 in London. For those contemplating moving to Big Data from often Relational Background
Applications of Big Data Analytics in BusinessesT.S. Lim
The document discusses big data and big data analytics. It begins with definitions of big data from various sources that emphasize the large volumes of structured and unstructured data. It then discusses key aspects of big data including the three Vs of volume, variety, and velocity. The document also provides examples of big data applications in various industries. It explains common analytical methods used in big data including linear regression, decision trees, and neural networks. Finally, it discusses popular tools and frameworks for big data analytics.
This document provides an overview of a presentation on big data and data science. It covers:
1. An introduction to key concepts in big data including architecture, Hadoop, sources of data, and definitions.
2. Details on common big data reference architectures from companies like IBM, Oracle, SAP, and open source technologies.
3. A discussion of how data science is disrupting various industries and the characteristics of firms using data science successfully.
4. Descriptions of machine learning techniques like segmentation, forecasting, and the overall reference architecture for machine learning involving data storage, signal extraction, and responding to insights.
Big data, Machine learning and the AuditorBharath Rao
This document discusses how auditors can use big data, machine learning, and analytics. It defines big data and machine learning, describing techniques like supervised vs. unsupervised learning. It provides examples of how auditors could use these approaches for risk management, fraud detection, process mining, compliance, and more. Specific use cases are outlined, like vendor collusion identification and predictive analytics for bad debts. Statistical methods like logistic regression that could support predictive analytics are also mentioned. The document suggests auditors are well-positioned to help companies implement big data and machine learning for assurance, automation, and controls.
Data science involves using industrial research techniques on a company's own data to develop advanced algorithms that provide a competitive advantage. Data engineering is a specialized form of software engineering focused on handling and processing data using skills in areas like structured and unstructured data storage, machine learning platforms, and predictive APIs. While data science and business intelligence overlap in using data analysis, statistics, and visualization, data science has a more scientific approach focused on the future rather than the past. Data-focused jobs are in high demand across many industries, especially technology, but some roles may become automated, increasing the value of skills like research and communication. Education options for these fields include academic programs, boot camps, and online classes.
A Seminar Presentation on Big Data for Students.
Big data refers to a process that is used when traditional data mining and handling techniques cannot uncover the insights and meaning of the underlying data. Data that is unstructured or time sensitive or simply very large cannot be processed by relational database engines. This type of data requires a different processing approach called big data, which uses massive parallelism on readily-available hardware.
This presentation is an Introduction to the importance of Data Analytics in Product Management. During this talk Etugo Nwokah, former Chief Product Officer for WellMatch, covered how to define Data Analytics why it should be a first class citizen in any software organization
This document provides tips for aspiring data scientists. It advises them to start by focusing on a topic that interests them and to clearly define their objectives and data collection process. It also recommends that they visualize their data, understand the context, look for additional insights, evaluate results, and find effective uses of the data. The document notes that data is becoming increasingly important in all industries and companies without data-savvy managers will be at a disadvantage.
This document provides an introduction to big data and analytics. It discusses the topics of data processing, big data, data science, and analytics and optimization. It then provides a historic perspective on data and describes the data processing lifecycle. It discusses aspects of data including metadata and master data. It also discusses different data scenarios and the processing of data in serial versus parallel formats. Finally, it discusses the skills needed for a data scientist including business and domain knowledge, statistical modeling, technology stacks, and more.
Data Science Applications | Data Science For Beginners | Data Science Trainin...Edureka!
** Data Science Certification using R: https://www.edureka.co/data-science **
This Edureka "Data Science Applications" PPT takes you through the various domains in which data science is being deployed today, along with some potential applications of this technology. The world today runs on data and this PPT shows exactly that.
Check out our Data Science Tutorial blog series: http://bit.ly/data-science-blogs
Check out our complete Youtube playlist here: http://bit.ly/data-science-playlist
Follow us to never miss an update in the future.
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Attend The Data Science Course in Bangalore From ExcelR. Practical Data Science Course in Bangalore Sessions With Assured Placement Support From Experienced Faculty. ExcelR Offers The Data Science Course in Bangalore.
Predictive Analytics - Big Data & Artificial IntelligenceManish Jain
Quick overview of the latest in big data and artificial intelligence. A lot of buzzwords being thrown around, hopefully this presentation will demystify many of the terms.
This document discusses the relationship between artificial intelligence (AI) and big data. It defines both AI and big data. AI is making computers do intelligent tasks like humans, while big data refers to large amounts of structured and unstructured data. The document explains that AI needs large amounts of data to replicate human intelligence and make intelligent decisions, just as human intelligence is built on experiences and data. It provides examples of how AI uses big data, such as Google's self-driving cars gathering sensor data to make driving decisions. The document also covers predictive analytics, unstructured data analysis, and data mining techniques like genetic algorithms and fuzzy logic.
Data science is a field that focuses on analyzing large amounts of data using tools and techniques to discover patterns and make business decisions. Data scientists utilize machine learning algorithms to develop predictive models from multiple sources of data in different formats. Data has become a valuable asset like oil in the 21st century that can help organizations improve decision making. The career is expected to grow exponentially and data scientists can earn more than average IT workers.
Big Data & Analytics (Conceptual and Practical Introduction)Yaman Hajja, Ph.D.
A 3-day interactive workshop for startups involve in Big Data & Analytics in Asia. Introduction to Big Data & Analytics concepts, and case studies in R Programming, Excel, Web APIs, and many more.
DOI: 10.13140/RG.2.2.10638.36162
Introduction to Data Science (Data Summit, 2017)Caserta
This document summarizes an introduction to data science presentation by Joe Caserta and Bill Walrond of Caserta Concepts. Caserta Concepts is an internationally recognized data innovation and engineering consulting firm. The agenda covers why data science is important, challenges of working with big data, governing big data, the data pyramid, what data scientists do, standards for data science, and a demonstration of data analysis. Popular machine learning algorithms like regression, decision trees, k-means clustering and collaborative filtering are also discussed.
This was first part of the presentation on "Road Map for Careers in Big Data" in Conjunction with Hortonworks/Aengus Rooney on 17th August 2016 in London. For those contemplating moving to Big Data from often Relational Background
Applications of Big Data Analytics in BusinessesT.S. Lim
The document discusses big data and big data analytics. It begins with definitions of big data from various sources that emphasize the large volumes of structured and unstructured data. It then discusses key aspects of big data including the three Vs of volume, variety, and velocity. The document also provides examples of big data applications in various industries. It explains common analytical methods used in big data including linear regression, decision trees, and neural networks. Finally, it discusses popular tools and frameworks for big data analytics.
This document provides an overview of a presentation on big data and data science. It covers:
1. An introduction to key concepts in big data including architecture, Hadoop, sources of data, and definitions.
2. Details on common big data reference architectures from companies like IBM, Oracle, SAP, and open source technologies.
3. A discussion of how data science is disrupting various industries and the characteristics of firms using data science successfully.
4. Descriptions of machine learning techniques like segmentation, forecasting, and the overall reference architecture for machine learning involving data storage, signal extraction, and responding to insights.
Big data, Machine learning and the AuditorBharath Rao
This document discusses how auditors can use big data, machine learning, and analytics. It defines big data and machine learning, describing techniques like supervised vs. unsupervised learning. It provides examples of how auditors could use these approaches for risk management, fraud detection, process mining, compliance, and more. Specific use cases are outlined, like vendor collusion identification and predictive analytics for bad debts. Statistical methods like logistic regression that could support predictive analytics are also mentioned. The document suggests auditors are well-positioned to help companies implement big data and machine learning for assurance, automation, and controls.
Data science involves using industrial research techniques on a company's own data to develop advanced algorithms that provide a competitive advantage. Data engineering is a specialized form of software engineering focused on handling and processing data using skills in areas like structured and unstructured data storage, machine learning platforms, and predictive APIs. While data science and business intelligence overlap in using data analysis, statistics, and visualization, data science has a more scientific approach focused on the future rather than the past. Data-focused jobs are in high demand across many industries, especially technology, but some roles may become automated, increasing the value of skills like research and communication. Education options for these fields include academic programs, boot camps, and online classes.
A Seminar Presentation on Big Data for Students.
Big data refers to a process that is used when traditional data mining and handling techniques cannot uncover the insights and meaning of the underlying data. Data that is unstructured or time sensitive or simply very large cannot be processed by relational database engines. This type of data requires a different processing approach called big data, which uses massive parallelism on readily-available hardware.
This presentation is an Introduction to the importance of Data Analytics in Product Management. During this talk Etugo Nwokah, former Chief Product Officer for WellMatch, covered how to define Data Analytics why it should be a first class citizen in any software organization
This document provides tips for aspiring data scientists. It advises them to start by focusing on a topic that interests them and to clearly define their objectives and data collection process. It also recommends that they visualize their data, understand the context, look for additional insights, evaluate results, and find effective uses of the data. The document notes that data is becoming increasingly important in all industries and companies without data-savvy managers will be at a disadvantage.
This document provides an introduction to big data and analytics. It discusses the topics of data processing, big data, data science, and analytics and optimization. It then provides a historic perspective on data and describes the data processing lifecycle. It discusses aspects of data including metadata and master data. It also discusses different data scenarios and the processing of data in serial versus parallel formats. Finally, it discusses the skills needed for a data scientist including business and domain knowledge, statistical modeling, technology stacks, and more.
Data Science Applications | Data Science For Beginners | Data Science Trainin...Edureka!
** Data Science Certification using R: https://www.edureka.co/data-science **
This Edureka "Data Science Applications" PPT takes you through the various domains in which data science is being deployed today, along with some potential applications of this technology. The world today runs on data and this PPT shows exactly that.
Check out our Data Science Tutorial blog series: http://bit.ly/data-science-blogs
Check out our complete Youtube playlist here: http://bit.ly/data-science-playlist
Follow us to never miss an update in the future.
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Attend The Data Science Course in Bangalore From ExcelR. Practical Data Science Course in Bangalore Sessions With Assured Placement Support From Experienced Faculty. ExcelR Offers The Data Science Course in Bangalore.
Predictive Analytics - Big Data & Artificial IntelligenceManish Jain
Quick overview of the latest in big data and artificial intelligence. A lot of buzzwords being thrown around, hopefully this presentation will demystify many of the terms.
This document discusses the relationship between artificial intelligence (AI) and big data. It defines both AI and big data. AI is making computers do intelligent tasks like humans, while big data refers to large amounts of structured and unstructured data. The document explains that AI needs large amounts of data to replicate human intelligence and make intelligent decisions, just as human intelligence is built on experiences and data. It provides examples of how AI uses big data, such as Google's self-driving cars gathering sensor data to make driving decisions. The document also covers predictive analytics, unstructured data analysis, and data mining techniques like genetic algorithms and fuzzy logic.
Data science is a field that focuses on analyzing large amounts of data using tools and techniques to discover patterns and make business decisions. Data scientists utilize machine learning algorithms to develop predictive models from multiple sources of data in different formats. Data has become a valuable asset like oil in the 21st century that can help organizations improve decision making. The career is expected to grow exponentially and data scientists can earn more than average IT workers.
Big Data & Analytics (Conceptual and Practical Introduction)Yaman Hajja, Ph.D.
A 3-day interactive workshop for startups involve in Big Data & Analytics in Asia. Introduction to Big Data & Analytics concepts, and case studies in R Programming, Excel, Web APIs, and many more.
DOI: 10.13140/RG.2.2.10638.36162
Introduction to Data Science (Data Summit, 2017)Caserta
This document summarizes an introduction to data science presentation by Joe Caserta and Bill Walrond of Caserta Concepts. Caserta Concepts is an internationally recognized data innovation and engineering consulting firm. The agenda covers why data science is important, challenges of working with big data, governing big data, the data pyramid, what data scientists do, standards for data science, and a demonstration of data analysis. Popular machine learning algorithms like regression, decision trees, k-means clustering and collaborative filtering are also discussed.
This document provides an overview of big data and discusses key concepts. It begins by defining big data and noting the increasing volume, velocity and variety of data being created. It then covers the big data landscape including storage models and technologies like Hadoop, analytics techniques like machine learning, and visualization. Finally, it discusses business uses cases and how big data is impacting industries and creating new business models through insights gained from data.
As machine learning has is permeating more and more industries and businesses, the need for audit professionals to provide assurance over machine learning is growing. Andrew's presentation will provide an audit-centric overview of machine learning and present a framework for how to begin auditing machine learning in your organization.
This document provides an overview of being a data science product manager. It discusses the speaker's journey becoming a PM, introduces data science applications in e-commerce, outlines the typical journey of building an AI/ML product, and discusses PM responsibilities. It also covers when to use AI/ML, includes a mini case study on recommendation engines, and discusses challenges including the non-deterministic nature of data science and lack of explainability in models.
MLOps is the process of taking machine learning models into production and maintaining and monitoring them. It addresses issues like lack of reproducibility, inability to identify new trends, and lack of scalability that can occur without proper processes. The machine learning lifecycle includes scoping a project, collecting and preparing data, developing and evaluating models, deploying models into production, and ongoing monitoring. MLOps aims to operationalize this lifecycle to ensure models can be deployed and updated efficiently and reliably at scale.
Overview of analytics and big data in practiceVivek Murugesan
Intended to give an overview of analytics and big data in practice. With set of industry use cases from different domains. Would be useful for someone who is trying to understand Analytics and Big Data.
A step towards machine learning at accionlabsChetan Khatri
This document provides an overview of machine learning including definitions of common techniques like supervised learning, unsupervised learning, and reinforcement learning. It discusses applications of machine learning across various domains like vision, natural language processing, and speech recognition. Additionally, it outlines machine learning life cycles and lists tools, technologies, and resources for learning and practicing machine learning.
Are We Generation AI? An Introduction to Applications, Benefits, and Challenges of AI for Small and Medium Sized Business. Presented at the WIN.fbg meeting in Fredericksburg, TX on April 11, 2023.
Discover Practical AI use cases in Customer Service! In this webinar, you will learn how to lower the time to first response and time to resolution to keep your SLAs intact, as well as about chatbots, ticket tagging, and urgency detection. We will also mention some technologies, such as text recognition and sentiment analysis.
My slides for my talk regarding machine learning and data science. Includes working examples with accompanying repo with reproducible code and data sets available.
Machine Learning: What Assurance Professionals Need to Know Andrew Clark
Machine learning has evolved past an esoteric technique worked on by academics and research institutes into a viable technology being deployed at many companies. Machine learning has been significantly changing the competitive landscape of business models worldwide, contributing to the demise of established business, such as Blockbuster, to creating entirely new businesses, such as algorithmic advertising. This presentation strives to address the questions of what assurance professionals need to know about this technology and how to provide assurance around machine learning implementations and its unique risks.
This document provides an overview of investing in AI-driven startups. It outlines Dr. Roy Lowrance's background working with machine learning systems and startups. It then lists 100 AI startups that have raised over $11.7 billion total. The agenda covers an overview of AI, machine learning and big data, the life cycle of AI projects, and sustainable competitive advantages for AI-based startups.
Introduction to machine learning and applications (1)Manjunath Sindagi
This document provides an introduction to machine learning including definitions, applications, and examples. It discusses the types of machine learning including supervised learning using examples of regression and classification. Unsupervised learning including clustering is also covered. The steps to solve a machine learning problem are outlined including feature selection, scaling, model selection, parameter selection, cost functions, gradient descent, and evaluation. Career opportunities in data science are discussed along with challenges such as data acquisition.
Presentation on developments in hiring and fintech for HKU Executive certific...Kok Tong (K.T.) Khoo
Slides for my guest speaker session at the HKU executive certificate in Internet Finance. Covering personal observations in startup markets and careers, Hong Kong vs Singapore, hiring trends and business models.
This document discusses moving from traditional business intelligence (BI) tools to adopting machine learning. It begins with an overview of common BI workflows and their limitations. It then provides introductions to machine learning, deep learning, and artificial intelligence. The machine learning pipeline is explained along with examples of adopting machine learning in products. Challenges of adopting machine learning are discussed as well as cost optimization strategies. Real world use cases are presented and open source options are mentioned.
Data Con LA 2022 - Demystifying the Art of Business Intelligence and Data Ana...Data Con LA
Brandon Wong, Lead Software Engineer, Academy of Motion Picture Arts and Sciences
Business Intelligence is a technology-driven process that analyzes data and forms conclusions to help assist workers to make informed business decisions. From collecting to cleaning, to morphing, to displaying we will address the pain points, tips, and tricks on how to navigate this process of converting data from raw material to a final product.
You'll learn:
From a high level, the process of bringing data from the "back" to the "front".
Tools and best practices for cleaning and displaying data.
Understanding the foundations of business intelligence to better execute on objectives.
The various ways of displaying data depends on circumstance.
Ai design sprint - Finance - Wealth managementChinmay Patel
Chinmay Patel presented an AI design sprint methodology. The methodology involves identifying a business problem, gathering and preparing relevant data, training and deploying a model, and maintaining/improving the model over time. As an example, Chinmay discussed how this process was used to build an automated claim resolution bot that can resolve claims within 3 seconds with no paperwork. The methodology was also proposed for a wealth management use case to perform user segmentation using clustering algorithms.
Book: Software Architecture and Decision-MakingSrinath Perera
Uncertainty is the leading cause of mistakes made by practicing software architects. The primary goal of architecture is to handle uncertainty arising from user cases as well as architectural techniques. The book discusses how to make architectural decisions and manage uncertainty. From the book, You will learn common problems while designing a system, a default solution for each, more complex alternatives, and 5Q & 7P (Five Questions and Seven Principles) that help you choose.
Book, https://amzn.to/3v1MfZX
Blog: http://tinyurl.com/swdmblog
Six min video - https://youtu.be/jtnuHvPWlYU
An Introduction to Blockchain for Finance ProfessionalsSrinath Perera
This document provides an introduction to blockchain technology for finance professionals. It discusses how blockchain can be used to establish trust in record-keeping in a similar way to double-entry bookkeeping. Blockchain acts as an indestructible append-only ledger that allows entries to be recorded but not edited, deleted, or repudiated. This enables trust to be established for processes like land registry and financial transactions. Blockchains can be public, with anyone able to participate, or private, restricted to certain participants. While not risk-free, blockchain ledgers can replace or reduce the need for intermediaries in areas like auditing, supply chain management, and financial transactions.
AI in the Real World: Challenges, and Risks and how to handle them?Srinath Perera
This document discusses challenges, risks, and how to handle them with AI in the real world. It covers:
- AI can perform tasks like driving a car faster and cheaper than humans, but can't fully explain how.
- Deploying and managing AI models at scale is complex, as is integrating models with user experiences. Bias and lack of transparency are also risks.
- When applying AI, such as in high-risk domains like medicine, it is important to audit models, gradually introduce them with trials, monitor outcomes, and find ways to identify and address errors or unfair impacts. With care and oversight, AI can be developed to help more people than it harms.
This document discusses how AI could shape future integrations. It begins by explaining different types of tasks that AI can perform, such as those that can be precisely explained versus those requiring examples and feedback to learn. The document then covers benefits of AI like speed, lower costs, and ability to learn and extrapolate. It discusses using AI for cost savings, competitive advantages, and new revenue streams through insights. Challenges of AI like lack of data and skilled professionals are presented along with risks such as bias, privacy issues, and how mistakes can be more harmful than for humans. Various use cases of AI in integration are explored such as enhancing inputs, security, and automatic integration. The document concludes that AI will create many new integration opportunities
The Role of Blockchain in Future IntegrationsSrinath Perera
We have critically evaluated blockchain-based integration use cases, their feasibility, and timelines. Emerging Technology Analysis Canvas (ETAC), a framework built to analyze emerging technologies, is the methodology of our study. Based on our analysis, we observe that blockchain can significantly impact integration use cases.
In our paper, we identify 30-plus blockchain-based use cases for integration and four architecture patterns. Notably, each use case we identified can be implemented using one of the architecture patterns. Furthermore, we also discuss challenges and risks posed by blockchains that would affect these architecture patterns.
Our webinar presents a critical analysis of serverless technology and our thoughts about its future. We use Emerging Technology Analysis Canvas (ETAC), a framework built to analyze emerging technologies, as the methodology of our study. Based on our analysis, we believe that serverless can significantly impact applications and software development workflows.
We’ve also made two further observations:
Limitations, such as tail latencies and cold starts, are not deal breakers for adoption. There are significant use cases that can work with existing serverless technologies despite these limitations.
We see a significant gap in required tooling and IDE support, best practices, and architecture blueprints. With proper tooling, it is possible to train existing enterprise developers to program with serverless. If proper tools are forthcoming, we believe serverless can cross the chasm in 3-5 years.
A detailed analysis can be found here: A Survey of Serverless: Status Quo and Future Directions. Join our webinar as we discuss this study, our conclusions, and evidence in detail.
1. Blockchain potential impact is real. If successful, Blockchain technologies can transform the way we live our day to day lives.
2. We believe technology is ready for limited applications in Digital Currency, Lightweight financial systems, Ledgers (of identity, ownership, status, and authority), Provenance (e.g. supply chains and other B2B scenarios) and Disintermediation, which we believe will happen in next three years.
3. However, with other use cases, blockchain faces significant challenges such as performance, irrevocability, need for regulation and lack of census mechanisms. These are hard problems and
4. It is not clear whether blockchain can sustain the current level of effort for extended period of 5+ years. There are many startups and they run the risk of running out of money before markets are ready. Failure of startups can inhibit further funding and investments.
5. Value and need of decentralization compared to centralized and semi-centralized alternatives is not clear.
A Visual Canvas for Judging New TechnologiesSrinath Perera
The document proposes an Emerging Technology Canvas (ETAC) framework to analyze emerging technologies. It is inspired by the Business Model Canvas and aims to provide a compact visual representation to capture the narrative around an emerging technology. The ETAC seeks to find the right questions to analyze technologies and understand their environment, impact, drivers, future potential and risks through a set of factors. It is presented as a tool to critically evaluate technologies and communicate insights about their adoption and development. Guidelines are provided on how to build an ETAC analysis and contribute insights to an online repository.
The talk discusses how analytics can attack privacy and what we can do about it. It discusses the legal responses (e.g. GDPR) as well technical responses ( differential privacy and homomorphic encryption).
The video is in https://www.facebook.com/eduscopelive/videos/314847475765297/ from 1.18.
Blockchain is often cited as one of the most impactful technology along with AI. It has attracted many startups, venture investments, and academic research. If successful, Blockchain technologies can transform the way, we live our day to day lives.
However, blockchain faces significant challenges such as performance, irrevocability, need for regulation and lack of census mechanisms. They are hard problems, and likely it will take at least 5-10 years to find answers to those problems.
Given the risk involved as well as the significant potential returns, we recommend a cautiously optimistic approach for blockchain with the focus on concrete use cases.
Today's Technology and Emerging Technology LandscapeSrinath Perera
We have seen the rise and fall of many technologies, some disappearing without a trace while others redefining the world. Collectively they have shaped our world beyond recognition. In this talk, Srinath will start with past technologies exploring their behavior. Then he will explore current middleware landscape, its composition, and relationships between different segments. He will discuss significant developments and discuss their future. Further, he will discuss emerging technologies, forces that shape them, and the promise of each technology, and finally, speculate about their evolution. You will walk away with knowledge on the evolution of middleware, the status quo, and discussion about how, at WSO2, we think those technologies will evolve.
Some died, some get by, but some have woven themselves to today's middleware so much that we do not notice them. The point I want to make is that not all emerging technologies are fads. Some are, and some are too early, like AI. But some are lasting.
The Rise of Streaming SQL and Evolution of Streaming ApplicationsSrinath Perera
Srinath Perera discusses the rise of streaming SQL and evolution of streaming applications. He covers what streaming is, how almost all new data is streaming, the streaming processing market, building streaming apps, the history of stream processing, why streaming SQL is useful, common solutions with stream processing, how stream processors are stateful and need high availability, how most are resource-intensive, the need for machine learning and advanced query authoring with stream processing. He then introduces WSO2 Stream Processor as a lightweight option for streaming applications.
Analytics and AI: The Good, the Bad and the UglySrinath Perera
Analytics let us question the data, which in effect questions the world around us. This let us understand, monitor, and shape the world. AI let us discover connections, predict the possible futures and automate tasks.
These twin technologies can change the world around us. On one hand, make us efficient, connected, and fulfilled. At the same time, the change of status quo can replace jobs, affect lives and build biases into our systems that can marginalize millions.
In this talk, we will discuss core ideas behind analytics and AI, their possible impact, both good and bad outcomes, and challenges.
The dawn of digital businesses is upon us, with reimagined business models that make the best use of digital technologies such as automation, analytics, integration and cloud. Digital businesses are efficient, continuously optimizing, proactive, flexible and are able to fully understand their customers. Analytics is a key technology that helps in doing so. It acts as the eyes and ears of the system and provides a holistic view on the past and present so that decision-makers can predict what will happen in the future. This webinar will explore
Why becoming a digital business is not a choice
The role of analytics in digital transformation with examples
How best to leverage state of the art analytics technology
SoC Keynote:The State of the Art in Integration TechnologySrinath Perera
This talk discusses Outline of the state of the art of Enterprise Software and how we get there, as I see it. Also second part describes Ballerina, a new programming language WSO2 has built for Enterprise Computing.
It is presented as a Keynote at 11th Symposium and Summer School On Service-Oriented Computing.
We are at the dawn of digital businesses, that are reimagined to make the best use of digital technologies such as automation, analytics, cloud, and integration. These businesses are efficient, continuously optimizing, proactive, flexible and able to understand customers in detail. A key part of a digital business is analytics: the eyes and ears of the system that tracks and provides a detailed view on what was and what is and lets decision makers predict what will be.
This session will explore how the WSO2 analytics platform
Plays a role in your digital transformation journey
Collects and analyzes data through batch, real-time, interactive and predictive processing technologies
Lets you communicate the results through dashboards
Brings together all analytics technologies into a single platform and user experience
This document provides guidance on conducting online research in 4 parts: 1) Get a feel for the topic by searching broadly and reading a variety of sources to understand different perspectives, 2) Select and refine relevant facts, 3) Make an argument by logically connecting facts, and 4) Write down the argument, distinguishing conclusions from opinions and citing sources. It emphasizes finding trustworthy sources, questioning biases and incentives, and thinking critically about one's own opinions through discussion with others. The goal of online research is finding truth by deriving conclusions from verified facts through logical reasoning and common sense.
Mieke Jans is a Manager at Deloitte Analytics Belgium. She learned about process mining from her PhD supervisor while she was collaborating with a large SAP-using company for her dissertation.
Mieke extended her research topic to investigate the data availability of process mining data in SAP and the new analysis possibilities that emerge from it. It took her 8-9 months to find the right data and prepare it for her process mining analysis. She needed insights from both process owners and IT experts. For example, one person knew exactly how the procurement process took place at the front end of SAP, and another person helped her with the structure of the SAP-tables. She then combined the knowledge of these different persons.
How iCode cybertech Helped Me Recover My Lost Fundsireneschmid345
I was devastated when I realized that I had fallen victim to an online fraud, losing a significant amount of money in the process. After countless hours of searching for a solution, I came across iCode cybertech. From the moment I reached out to their team, I felt a sense of hope that I can recommend iCode Cybertech enough for anyone who has faced similar challenges. Their commitment to helping clients and their exceptional service truly set them apart. Thank you, iCode cybertech, for turning my situation around!
[email protected]
AI Competitor Analysis: How to Monitor and Outperform Your CompetitorsContify
AI competitor analysis helps businesses watch and understand what their competitors are doing. Using smart competitor intelligence tools, you can track their moves, learn from their strategies, and find ways to do better. Stay smart, act fast, and grow your business with the power of AI insights.
For more information please visit here https://www.contify.com/
This comprehensive Data Science course is designed to equip learners with the essential skills and knowledge required to analyze, interpret, and visualize complex data. Covering both theoretical concepts and practical applications, the course introduces tools and techniques used in the data science field, such as Python programming, data wrangling, statistical analysis, machine learning, and data visualization.
Just-in-time: Repetitive production system in which processing and movement of materials and goods occur just as they are needed, usually in small batches
JIT is characteristic of lean production systems
JIT operates with very little “fat”
1. Data Science Use Cases in
the Enterprise
Srinath Perera
Chief Architect, WSO2, Apache Member
2. Context: Understanding
Enterprise (ROI)
● It is about Money: long-term Money.
○ If you are looking to make a million once, sometimes,
you can get away with exploitation.
○ If you are looking to make a billion every year, you
have to care about customers, brand, employees as
well as the environment you are operating in
○ E.g., Indra Nooyi and her effort to move Pepsi to
healthy food.
● It is a Strategic environment where enterprises
compete.
○ “If you know the enemy and know yourself, you need not fear
the result of a hundred battles. ”
― Sun Tzu, The Art of War.
3. Context: Highly valued
Outcomes
● Efficiency, Savings
● Improving Customer Experience
● Finding new markets,
understanding markets
● Forecasts, Prediction
● Automation and Decision Support
I skate to where the puck is
going to be, not where it has
been. ---Wayne Gretzky
4. ● Examples
○ The effort by the US to use sensor and data analysis to stop
infiltration through Ho Chi Minh Trail in 70s
○ Even Nate Silver got Trump's victory wrong
● Reasons
○ History is not always representative of the future (e.g., Trump
Elections)
○ Complex systems ( highly interconnected systems where one
or few players can significantly change the outcomes)
○ Highly competitive situations such as stock Markets
■ Predictable at stable times, but not with shock
○ Average is affected dramatically by rare events (e,g, Covid)
■ Data can determine "average" outcomes with great
accuracy
○ Not enough data or data do not capture critical aspects
Nevermind the Press, Data Science does not always work
5. Use Cases @ Enterprise
● Efficiency, Self Awareness, and Forecasts
● Optimizing the sales funnel
● Predictive Maintenance
● Improving Customer Experience
● Product Use cases from a real-world iPaaS
● Finding new markets, understanding markets, Competitor
Analysis
● https://sparktoro.com/ - Instantly discover what your
audience reads, watches, listens to, and follows.
● Automate mundane tasks and let people focus on what
they are good at
● Automation and Task Assistant Systems
● Decision support systems
Often needs Explainability too
6. Efficiency: Optimize the Sales Funnel
● Each enterprise has a funnel
like this ( names may be
different)
● KPIs support decisions
● Examples:
○ conversion rates, dropoff - to find
bottlenecks
○ cost per conversion - find
activities that work well
○ Time spend on each stage
○ Forecasts
○ A/B testing optimizes
7. Efficiency: Predictive
Maintenance
● Often breakdowns have high costs
● We do preventive maintenance to
avoid that, but it leaves significant
money on the table
● Use telemetry data to predict
breakdowns
● We need to manage risk against
false negatives (e.g., cost to give
customer 100$)
8. Efficiency: Churn Prediction
● Even small churn compounds
significantly to reduce topline, and create
negative word of mouth.
● How is the user using the product?
● Has he given up?
● Are there complaints?
● Is there anything we can do if we know
before?
Need to think through the full story -
Ask “so what” until you see $$
10. Choreo Use Cases and Challenges
● Can collect data about everything, clicks,
messages, logs etc
● The focus is using AI to improve user
experience
● The system will have 10s of thousands of users
○ We can’t run a model per user
● Some use cases have limited data
● The specific user would not have enough data
initially, so we have a cold start problem
● Some use cases require personalization
11. User Experience: Forecasting Performance
● Performance feedback while
you write code
● API, service, database calls
dominate performance
● Use historical data about each
API, service, database call and
fit Machine Learning models
● Use queuing theory to model
the throughput and latency
12. Getting a Model to Production is Complicated
● Data Collection
● Model training
● Model deployment and
integrating the model into the
user experience
○ Acting on results
● Getting user feedback
● Evaluating and improving
models
13. User Experience: Automatic Data Mapping
● Programming with APIs
need us to map data
between two API calls (
and two systems)
● Automatic data
mapping suggest
mapping between two
data types
● It can maps data types
it has never seen
14. User Experience: Anomaly and Root Cause Prediction
● Detecting Performance anomalies in
the system
● The goal is to detect and performance
problems and notify the users and
supporting them in troubleshooting
● We started with several states of the art
papers and eventually beat them
○ 90% precision and 50% recall vs. 98% vs.
81% recall
● Working on attributing anomalies to
parts of the system and providing root
cause predictions
42
16. Automation: Extracting information from Images/ Video
● Vidado.ai Using OCR to digitize Data RPA
does not work well with paper
● Icetana.com - decision support for video
surveillance
● www.dataminr.com detects high impact
events from public data
○ E.g., Brand risk, disease outbreaks, potential
new stories
17. Automation: Competitive Adjustments
● Common use cases
are adjusting the price
● This leads to curious
cases when bots are
on both sides
A good rule of thumb is to remember AI vs. AI does not work well.
18. Automation: Automate Mundane Tasks
● Works on top
salesforce
● Suggest next Action
● Provides templates
for actions
● Full context, connect
all information
● Benchmark
performance
19. Parting Thoughts
● If you plan to solve organizational problems
with data science, you need to understand
how it works and speak their language.
● Make sure there is enough data
● Think through the full lifecycle, including
economics (e.g., Choreo) and explain
● Model deployment, evaluation, integration to
customer, and evolution is complex
● Harder to build per user custom models,
better if you can create value against existing
data models and integrate as SaaS
Learn to see where
Data Science works,
but learn to see where
it does not also!!