SDS Deep Learning For Spatial Application
SDS Deep Learning For Spatial Application
Computer Speech
Vision Machine Recognition
Learning
Random Forest
Logistic Regression
Linear regression
SVM
Deep
Learning
Convolutional
Neural Networks
Mathematical Modelling of
Neural Network
Mathematical Modeling
• Creates a mathematical representation of some
phenomenon to better understand it.
• Matches observation with symbolic
representation.
• Informs theory and explanation.
𝑌 = 𝐵𝑖𝑎𝑠 + 𝑊. 𝑋 +error
Sejarah Jaringan Saraf Tiruan
Ide jaringan saraf dimulai sebagai model dari fungsi neuron di otak.
Pada tahun 1943, ahli neurofisiologi Warren McCulloch dan
matematikawan Walter Pitts menggambarkannya dengan sirkuit
listrik sederhana.
Donald Hebb mengambil gagasan tersebut dan menuliskan dalam
bukunya, The Organisation of Behavior (1949).
Dua konsep utama yang merupakan pendahulu dari Neural
Networks adalah:
'Threshold Logic' - mengubah input kontinu menjadi output diskrit
'Hebbian Learning' - model pembelajaran berdasarkan plastisitas
saraf, yang dikemukakan oleh Donald Hebb dalam bukunya "The
Organisation of Behavior" yang sering diringkas dengan frasa:
"Sel yang menyala bersama, menyatu.
Keduanya diusulkan pada tahun 1940-an. Pada tahun 1950-an,
peneliti mulai mencoba menerjemahkan jaringan ini ke sistem
komputasi.
McCulloch-Pitt Neuron
-Boolean inputs
-Boolean output
-Be able to model logic function such as OR, AND,
NOT
-Does not have ability to learn so threshold b need
to be adjust analytically to fit the output
Pada tahun 1950-an tersebut Frank Rosenblatt, seorang psikolog di
Cornell, sedang berusaha memahami sistem keputusan yang relatif
lebih sederhana yang ada di mata lalat.
Dalam upaya untuk memahami dan mengukur proses ini, dia
mengusulkan gagasan Perceptron pada tahun 1958, serta
menyebutnya Mark I Perceptron.
Gagasan tersebut berupa sistem dengan hubungan input output
sederhana beserta bobot, dimodelkan pada neuron McCulloch-
Pitts.
Pada tahun 1969, Marvin Minsky dan Seymour Papert menerbitkan
Perceptrons - teks bersejarah yang akan mengubah jalannya
penelitian kecerdasan buatan selama beberapa dekade.Dalam teks
tersebut, Minsky dan Papert membuktikan bahwa satu perceptron -
kakek dari unit komputasi yang menyusun jaringan saraf modern -
tidak mampu mempelajari fungsi eksklusif-atau (alias XOR).
Pada tahun 1986, Hinton, Rumelhart, dan Williams menerbitkan
sebuah makalah “Learning representations by back-propagating
errors”, memperkenalkan konsep backpropagation dan hidden
layers, oleh sebab itu dapat dikatakan mereka melahirkan
Multilayer Perceptrons (MLPs)
Backpropagation, prosedur untuk menyesuaikan bobot berulang kali
untuk meminimalkan perbedaan antara keluaran aktual dan prediksi
Hidden layers, yang merupakan node neuron yang ditumpuk di antara
input dan output, memungkinkan jaringan neural mempelajari fitur yang
lebih rumit (seperti logika XOR)
JST Dari Sudut Pandang Statistika
“There has been much publicity about the ability of artificial neural
networks to learn and generalize. In fact, the most commonly used
artificial neural networks, called multilayer perceptrons, are nothing
more than nonlinear regression and discriminant models that can be
implemented with standard statistical software.”
class A
class A
Anomaly Detection
Sequence labeling
http://mbjoseph.github.io/2013/11/27/measure.html
…
What is a Neural Net?
Structure: input-processing-output
Mimic neuronal signal firing structure of brain with
computational processing units
Source: https://www.slideshare.net/ThomasDaSilvaPaula/an-introduction-to-machine-learning-and-a-little-bit-of-deep-learning,
http://cs231n.github.io/convolutional-networks/
Artificial Neural Network
Weights
Activation functions
How do we train?
https://www.xenonstack.com/blog/static/public/uploads/media/machine-learning-vs-deep-learning.png
Conceptual Definition:
Deep learning is a computer program that
can
identify what something is
Technical Definition:
Deep learning is a class of machine learning
algorithms in the form of a neural network that
uses a cascade of layers (tiers) of processing
units to extract features from data and make
predictive guesses about new data
Source: Extending Jann LeCun, http://spectrum.ieee.org/automaton/robotics/artificial-intelligence/facebook-ai-director-yann-lecun-
on-deep-learning
Why is DL useful?
o Manually designed features are often over-specified, incomplete and
take a long time to design and validate
o Learned Features are easy to adapt, fast to learn
o Deep learning provides a very flexible, (almost?) universal, learnable
framework for representing world, visual and linguistic information.
o Can learn both unsupervised and supervised
o Effective end-to-end joint system learning
o Utilize large amounts of training data
Source: Yann LeCun, CVPR 2015 keynote (Computer Vision ), "What's wrong with Deep Learning" http://t.co/nPFlPZzMEJ
Convolutional Neural
Networks (CNNs)
Convolutional Neural Network (CNN) is one of the latest developments of artificial
neural networks inspired by human neural networks and commonly used in image data to
detect and recognize an object in an image. CNN consists of neurons that have weights,
biases and activation functions.
Convolutional
Input matrix 3x3 filter
http://deeplearning.stanford.edu/wiki/index.php/Feature_extraction_using_convolution
Convolutional Layer
Convolutional layer section performs convolution operations by using linear
filters on local areas. This layer is the first step to receive an image feeded into
the deep learning architecture. This layer is a filtered with specific filter matrix
with certain length (pixel), width (pixel) and dimension according to image
channel/band of the feeded data. These filters will shift throughout the image
according stride parameter. Filter direction is from left to the right and from top
to the bottom of matrix image. This shift will do a "dot" operation between the
input and the value of the filter so that it will produce an output called the
activation map or feature map. Figure 4 shows the convolution process in the
convolution layer and Figure 5 is how to calculate the convolution value.
Pooling
Layer
Main CNN idea for text:
Compute vectors for n-grams and group them afterwards
max pool
2x2 filters
and stride 2
https://shafeentejani.github.io/assets/images/pooling.gif
Fully Connected
•Fully connected layer takes input from the output pooling layer in the form of a
feature map. The feature map is still in the form of a multidimensional array so it
will reshape the feature map and generate n-dimensional vectors where n is the
number of output classes that the program must choose. For example, the layer
consists of 500 neurons, softmax will be applied which returns the list of greatest
probabilities for each of the 10 class labels as the final classification of the
network. Figure7 shows the process in the fully connected layer.
CNN
Architecture
Evaluation metrics of
deep learning training
𝑇𝑃+𝑇𝑁
𝐴𝑐𝑐𝑢𝑟𝑎𝑐𝑦 =
𝑇𝑃+𝑇𝑁+𝐹𝑃+𝐹𝑁
Where
TP = True Positives,
TN = True Negatives,
FP = False Positives, and
FN = False Negatives.
How does the neural net actually learn?
Structural system based on cascading layers of
neurons with variable parameters: weight and bias
Vary the weights
and biases to see if
a better outcome is
obtained
Repeat until the net
correctly classifies
the data
Source: http://neuralnetworksanddeeplearning.com/chap2.html
Deep Learning frameworks and libraries
Source: http://www.infoworld.com/article/3163525/analytics/review-the-best-frameworks-for-machine-learning-and-deep-
learning.html#tk.ifw-ifwsb
Hardware
Advances in chip design
GPU chips (graphics processing unit):
3D graphics cards designed to do fast
matrix multiplication
Google TPU chip (tensor processing
unit): custom ASICs for machine
learning, used in AlphaGo
TPUs process matrix Google TPU chip (Tensor
multiplications without storing Processing Unit), 2016
https://tensorflow.rstudio.com/gallery/
What are
layers?
• Data transformation functions parameterized by weights
• A layer is a geometric transformation function on the data that goes
through it (transformations must be differentiable for stochastic gradient
descent)
• Weights determine the data transformation behavior of a layer
R examples in
the gallery
• https://tensorflow.rstudio.com/gallery/
40
Studi Kasus
Penilaian Potensi Bahaya Longsor Menggunakan ML
X variables
B1="slope.tif"
B2="ndvi.tif"
B3="landcover.tif"
B4="elevation.tif"
B5="curvature.tif"
Y variables
Landslide occurence