Lecture 2 Final
Lecture 2 Final
2
Cognition
Definition
Examples of Processes
• Understanding spoken or written language
• Solving problems and making decisions
• Recognizing patterns and interpreting emotions
3
Cognition
Goal
Key Features
4
Cognition Vs Cognitive Computing
Aspect Cognition Cognitive Computing
5
Human reasoning, perception, AI-driven systems like IBM Watson: answering questions
Examples
memory. in Jeopardy or assisting in medical diagnosis, chatbots.
Cognitive Computing - Goal
Goal
6
Human Intelligence vs AI
7
AI vs Cognitive Computing
Aspect AI Cognitive Computing
Definition Focuses on building systems that replace Focuses on building systems that
human intelligence for automating tasks augment/assist human intelligence, helping
and decision-making. humans make better decisions.
Technology Machine Learning, Natural Language ML, Natural Language Processing, Big Data
Processing, Neural Networks, Deep Analytics, Neural Networks, Deep Learning,
Learning Sentiment Analysis
Interaction
Minimal or none High
with humans
8
AI vs Cognitive Computing
Aspect AI Cognitive Computing
Applications Autonomous systems, such as robotics, Decision-support systems in industries like
self-driving cars (Tesla), and predictive healthcare, finance, and education.
analytics.
- IBM Watson assisting doctors with
- AI Powered chatbots like Siri, Alexa diagnoses
- Financial Decision Support Systems
- Customer Service Improvement through
NLP
9
Foundation of Cognitive Computing
• Natural Language Processing (NLP)
• Machine Learning (ML)
• Big Data Analytics
10
Natural Language Processing (NLP)
Definition
Key Techniques
Breaking down text into smaller, manageable units, called tokens.
• Tokenization
Word Tokenization:
Sentence Tokenization:
11
Natural Language Processing (NLP)
Definition
Key Techniques
Breaking down text into smaller, manageable units, called tokens.
Analyzing the grammatical structure of a sentence to
understand the meaning
Word Tokenization:
For the sentence "The cat sits on the mat.", parsing identifies:
• Parsing Subject: "The cat"
Verb: "sits"
Sentence Tokenization:
Prepositional phrase: "on the mat"
12
Natural Language Processing (NLP)
Definition
Key Techniques
Breaking down text into smaller, manageable units, called tokens.
Analyzing the grammatical structure of a sentence to
Also known asthe
understand opinion mining
meaning
Word Tokenization:
Used
Forto
thedetermine thecat
sentence "The sentiment
sits on the or emotion
mat.", parsingexpressed
identifies: in a
piece"The
Subject: of text
cat"
Verb: "sits"
Sentence Tokenization:
Prepositional phrase: "on the mat"
• Sentiment analysis
13
Natural Language Processing (NLP)
Applications
14
Machine Learning (ML)
Definition
Types
• Supervised
• Unsupervised, and
• Reinforcement Learning
15
Machine Learning (ML)
Key Algorithms
• A field that involves examining large and complex data sets to uncover
hidden patterns, correlations, and insights, enabling data-driven
decision-making.
Key Techniques
• Data mining, data visualization, predictive analytics, clustering, and real-time
processing.
17
Big Data Analytics
Applications
18
Big Data Analytics
Structured data
• Data that can be processed, stored, and retrieved in a fixed format.
• It refers to highly organized information that can be readily and seamlessly stored
and accessed from a database by simple search engine algorithms.
• Example-The employee table in a company database will be structured as the
employee details, their job positions, their salaries, etc., will be present in an organized
manner.
19
Big Data Analytics
Unstructured data
• Refers to the data that lack any specific form or
structure whatsoever.
• This makes it very difficult and time-consuming
analyze unstructured data.
• Stored in data lakes, NoSQL (Not Only Structured
Query Language) databases, or cloud-based
systems.
• Free-form or undefined format (e.g., text, images,
videos, audio). 20
Big Data Analytics
Semi-structured data
• Pertains to the data containing both the formats
mentioned above, that is, structured and unstructured data.
• To be precise, it refers to the data that although has
not been classified under a particular repository
(database), yet contains vital information or tags that
segregate individual elements within the data.
• Example- Email
• The written content of an email is unstructured,
whereas the there are some inherent structure to the
information in each email, such as the sender’s name, 21
23
All definitions have 5 common characteristics
Characteristics of Big Data – 5Vs
24
• Velocity is the speed at which data is being created
in real time.
• Data is being generated extremely fast, in a process
that never stops
• Near or real-time streaming, local, and cloud-based
technologies can process information very quickly
25
• Volume is the scale of the data, or the increase in the
amount of data stored.
• Drivers of volume are the increase in data sources,
higher resolution sensors, and scalable infrastructure
• A large amount of data is stored in data warehouses.
26
• Variety is the diversity of the data.
• Structured data fits neatly into rows and columns, in relational
databases while unstructured data is not organized in a pre-defined
way, like Tweets, blog posts, pictures, numbers, and video.
• Variety also reflects that data comes from different sources,
machines, people, and processes, both internal and external to
organizations.
• Drivers are mobile technologies, social media, wearable
technologies, geo technologies, video, and many, many more.
Machines
People,
27
Processes
• Veracity is the quality and origin of data, and its conformity to facts
and accuracy.
• Attributes include consistency, completeness, integrity, and
ambiguity.
• Drivers include cost and the need for traceability.
• With the large amount of data available, the debate rages on about
the accuracy of data in the digital age.
• Is the information real, or is it false?
28
• Value is our ability and need to turn data into value.
• Value isn't just profit.
• It may have medical or social benefits, as well as
customer, employee, or personal satisfaction.
• The main reason that people invest time to understand
Big Data is to derive value from it.
29
Fundamental Principles of Cognitive Computing
Learn
31
Fundamental Principles of Cognitive Computing
32
Fundamental Principles of Cognitive Computing
• The hypothesis is defined as the supposition or proposed
explanation based on insufficient evidence or assumptions
• It is just a guess based on some known facts but has not yet
been proven.
• A good hypothesis is testable, which results in either true or
false.
• Example:
Generate • Some scientist claims that ultraviolet (UV) light can damage the eyes then it
may also cause blindness.
Hypotheses • In this example, a scientist just claims that UV rays are harmful to the eyes,
but we assume they may cause blindness.
• However, it may or may not be possible. Hence, these types of assumptions
are called a hypothesis.
33
Fundamental Principles of Cognitive Computing
• Generate Hypothesis-
• A cognitive system assumes that there is not a single correct answer.
• The most appropriate answer is based on the data itself.
• Therefore, a cognitive system is probabilistic.
• A hypothesis is a candidate explanation for some of the data already
understood.
• A cognitive system uses the data to train, test, or score a hypothesis.
Generate
Hypotheses
34