IntroAI-Slide01
IntroAI-Slide01
AI
TAE-HOON KIM
SLIDE 01
Industrial Revolutions
Industrial Revolutions
01 02 03 04
Environment
Sensors
Percepts
Agent
dictate techniques for selecting rational ?
actions Actuators
Actions
Brief history of AI
Intelligence
Capacity for logic, understanding, self-awareness, emotional
knowledge, reasoning, planning, creativity, critical thinking, and
problem-solving
It should be able to
◦ Interact with the real world
◦ Perform Reasoning and planning
◦ Learn and adapt
What Can AI Do?
Play a decent game of Jeopardy?
Win against any human at chess?
Win against the best humans at Go?
Play a decent game of tennis?
Grab a particular cup and put it on a shelf?
Unload any dishwasher in any home?
Drive safely along the highway?
Drive safely along Calumet Avenue?
Buy a week's worth of groceries on the web?
Buy a week's worth of groceries at Whole Food Market?
Discover and prove a new mathematical theorem?
Perform a surgical operation?
Translate spoken Korean into spoken English in real time?
Write an intentionally funny story?
Deep Blue (IBM) – Game Agent
Classic Moment: May, '97: Deep Blue vs. Kasparov
◦ First match won against world champion
◦ “Intelligent creative” play
◦ 200 million board positions per second
◦ Humans understood 99.9 of Deep Blue's moves
◦ Can do about the same now with a PC cluster
1996: Kasparov Beats Deep Blue
“I could feel --- I could smell --- a new kind of intelligence across
the table.”
1997: Deep Blue Beats Kasparov
“Deep Blue hasn't proven anything.”
Search problem formulation
States – environment
Initial state – the agent starts in
Actions – available actions to the agent
Transition model – returns the state resulting from action
Goal states – where the agent stops
Action cost – numeric cost of applying action in given state
Example – Sokoban puzzle
States
◦ State description specifies the location of each of the tiles
Initial state
◦ Any state can be designated as the initial state
Actions
◦ Physically, we move tiles
◦ In this problem, possible move of empty box
◦ (L, R, U, D), (R, U, D), (L, U, D), (R, U), (R, D), (L, U), (L, D)
Transition model
◦ Maps a state and action to a resulting state
Goal state
Action cost → Each action cost is 1
State Space Graphs
State space graph: A mathematical
representation of a search problem
◦ Nodes are (abstracted) world configurations
◦ Arcs represent successors (action results)
◦ The goal test is a set of goal nodes (maybe only one)
S’1
Random
S action a
S’2
Markov Decision Processes
An MDP is defined by: M = (S,A,T,R)
◦ A set of states s S
◦ A set of actions a A
◦ A transition function T(s, a, s’)
◦ Probability that a from s leads to s’, i.e., P(s’| s, a)
◦ Also called the model or the dynamics
◦ A reward function R(s, a, s’)
◦ Sometimes just R(s) or R(s’) for simplicity
◦ A start state
◦ Maybe a terminal state
Double-Bandit MDP
Actions: Blue, Red
No discount
States: Win, Lose 0.25 $0 10 time steps
Both states have
the same value
0.75
$2
W 0.25 L
$0
$1 $1
0.75 $2
1.0 1.0
Online Planning
Rules changed! Red’s win chance is different.
?? $0
??
$2
W ?? L
$0
$1 $1
?? $2
? ??
Reinforcement Learning
Still assume a Markov decision process (MDP):
◦ A set of states s S
◦ A set of actions (per state) A
◦ A model T(s,a,s’)
◦ A reward function R(s,a,s’)
35
AlphaStar
Robotics
Robotics
◦ Part mech. eng.
◦ Part AI
◦ Reality much harder than
simulations!
Technologies
◦ Vehicles
◦ Rescue
◦ Help in the home
◦ Lots of automation…
Images from UC Berkeley, Boston Dynamics, RoboCup, Google
Human-aid Interaction
Limitations
Narrowed models
Requires data and supervision
Can’t think like human
Safety, Ethics, and Privacy issues
Security
Cost and maintenance
Etc.
Utility
Clear utility function Not so clear utility function
Tools for Predictions & Decisions
Data
Definition
◦ Individual facts, statistic, or items of information
42
Data explosion
Moore’s Law
Digital storage becomes larger, cheaper, and faster with each
successive year
Since 1986, the amount of data storage dramatically 8000
increases 7000
2000
◦ 2014 5000 EB
1000
◦ 2020 6800 EB
0
1980 1985 1990 1995 2000 2005 2010 2015 2020 2025
43
Data
Data Age
Data sources
◦ New York Stock exchange – 4-5 terabytes per day
◦ Facebook – more than 240 billion photos
◦ Cloud
◦ IoT …
44
Big data
RDBMS for big data
Big data
◦ Larger, more complex data set
◦ New data sources
◦ Semi or unstructured data
45
Hadoop history
Started in 2002
◦ Doug Cutting and Mike Cafarella worked on Apache Nutch search engine project
funded by Yahoo
◦ They found that traditional database is too expensive for search engine to store
large volume of data
In 2003
◦ Google published GFS (Google File System) paper
Doug Cutting
In 2007
◦ Successfully tested Hadoop on 1000 node cluster
In 2008
◦ Yahoo released Hadoop as an open-source project to Apache Software
Foundation
◦ Successfully tested on 4000 node cluster
◦ Hadoop won Terabyte Sort Benchmark (209 seconds)
46
Reference: https://www.geeksforgeeks.org/hadoop-history-or-evolution/
Hadoop
Traditional approach
◦ Use local storage to store and access the data
◦ As the data size bigger, this approach is very complex and expensive
Hadoop approach
◦ In stead of local machines to fetch the data, it uses multiple nodes
◦ Each node holds part of the data
47
Machine Learning
Up until now: use a model to make optimal decisions
Artificial Machine
Intelligence Learning Deep
Learning
Simulation of human Learning describes an
intelligence in machines to automatic search Extract
mimic human behavior process for better patterns
representations
Deep Learning
Reference: https://towardsdatascience.com/why-deep-learning-is-needed-over-traditional-machine-learning-1b6a99177063
Why Deep Learning?
Manually feature extraction may easily cause over-specified, incomplete
and take a long time to design and validate
Can we find appropriate features from the data directly?
Deep learning provides a very flexible and universal
Works both unsupervised and supervised
Effective end-to-end joint system learning
Utilize large amounts of training data
Traditional ML vs. Deep Learning
Neural networks
Artificial network consists of artificial neurons or nodes
Each neurons or nodes are processing units
They are communicating each other with a weights
Each unit performs a relatively simple job
◦ Test: take input and compute an output propagating to next unit
◦ Learning: adjusting the weights
Neural Network
Inputs are feature values
Each feature has a weight
x1 w1
Sum and apply activation function
w2
Generate output x2 ∫ 𝑦ො
wn
…
xn
https://medium.com/ml-research-lab/under-fitting-over-fitting-and-its-solution-dc6191e34250