Lecture 1 Parallel and Scalable Machine Learning by HPC Morris Riedel
Lecture 1 Parallel and Scalable Machine Learning by HPC Morris Riedel
Selected Facts
One of EU largest
inter-disciplinary
research centres
(~5000 employees)
Special expertise
Physics
Material sciences
Nanotechnology
Neuroscience &
Medicine
Information technology
(HPC, Big Data, Quantum
Computing, Clouds, etc.)
Artificial Intelligence (AI) [25] Juelich Supercomputing Centre [24] Helmholtz Association Web Page
Domain-specific
Simulation Labs
SDLs
Cross-
Sectional
Team Deep PADC
Learning
Cross-Sectional Teams Data Life Cycle Labs Exascale co-Design
Selected Facts
Ranked among the top 200
universities in the world
(by Times Higher Education)
~2900 students at the
SENS school
Long collaboration with
Forschungszentrum Juelich
~350 MS students
~150 doctoral students.
Many foreign &
Erasmus students
English courses
[19] University of Iceland Web page
Community Building
Classification
Machine learning methods can be roughly categorized in classification, clustering, or regression augmented with various techniques for data
exploration, selection, or reduction – despite the momentum of deep learning, traditional machine learning algorithms are still widely relevant today
The basic meaning of learning is ‘to use a set of observations to uncover an underlying process‘ [14] Image sources: Species Iris Group of
The three different learning approaches are supervised, unsupervised, and reinforcement learning North America Database, www.signa.org
Supervised Learning
Majority of methods follow this
approach in this course
Example: credit card approval based
on previous customer applications
Unsupervised Learning
Often applied before other learning higher level data representation
Example: Coin recognition in vending [15] A.C. Cheng et al., ‘InstaNAS:
Instance-aware Neural
machine based on weight and size Architecture Search’, 2018
Reinforcement Learning
Typical ‘human way‘ of learning
Example: Toddler tries to touch a hot cup of tea (again and again)
Day 3 offers details about unsupervised learning with examples & Day 3 offers also a short introduction to deep reinforcement learning
Lecture 1 – Parallal and Scalable Machine Learning by HPC 14 / 50
Machine Learning Prerequisites & Challenges
Supervised learning approaches fits a model that related the response to the predictors
Supervised learning approaches are used in classification algorithms such as SVMs
Supervised learning works with data = [input, correct output]
Lecture 2 offers an example of using supervised learning with a known dataset & simple learning model to understand basic concepts
Lecture 1 – Parallal and Scalable Machine Learning by HPC 16 / 50
Simple Application Example: Classification of a Flower
(what type of flower is this?)
Lecture 2 offers an example of using supervised learning with a known dataset & simple learning model to understand basic concepts
Lecture 1 – Parallal and Scalable Machine Learning by HPC 17 / 50
A Simple Linear Learning Model – The Perceptron
(training data)
(modelled as
bias term)
(activation
function,
d
+1 or -1) (representing the threshold)
(the signal) (dimension of features)
Lecture 2 offers an example of using supervised learning with a known dataset & simple learning model to understand basic concepts
Lecture 1 – Parallal and Scalable Machine Learning by HPC 18 / 50
From Simple Perceptron to Innovative Deep Learning Techniques
Day 2 offers a more detailed introduction to Deep Learning Techniques with examples and Convolutional Neural Networks (CNNs)
Lecture 1 – Parallal and Scalable Machine Learning by HPC 19 / 50
Deep Learning Technique Example – Convolutional Neural Networks (CNNs)
[5] A. Rosebrock
Day 2 offers a more detailed introduction to Deep Learning Techniques with examples and Convolutional Neural Networks (CNNs)
Lecture 1 – Parallal and Scalable Machine Learning by HPC 20 / 50
Complex Relationships: ML & DL vs. HPC/Clouds & Big Data
Computing
High Performance
Training Computing & Cloud
Model Performance / Accuracy
Time
Large Deep Learning Networks Computing
‘small datasets‘
manual feature
engineering‘ Medium Deep Learning Networks
changes the
ordering
Small Neural Networks
[7] NVIDIA
image analysis
Combination: Start-up Example of my research group
Impact in AI & HPC
in industry & science [11] C. Bodenstein & M. Riedel et al., Automated Soccer Scence Tracking using Deep Neural Networks
High Performance Computing (HPC) is based on computing resources that enable the efficient use of parallel computing techniques
through specific support with dedicated hardware such as high performance cpu/core interconnections.
HPC
(network connection
very important & costly)
High Throughput Computing (HTC) is based on commonly available computing resources such as commodity PCs and small clusters that
enable the execution of ‘farming jobs’ without providing a high performance interconnection between the cpu/cores.
(network connection
less important)
HTC
This course is using HPC resources while the general techniques and algorithms can also work on HTC (e.g. Apache Spark, etc.)
Lecture 1 – Parallal and Scalable Machine Learning by HPC 26 / 50
Partnership for Advanced Computing in Europe (PRACE)
Basic Facts
HPC-driven infrastructure
An international not-for-profit
association under Belgien
law (with its seat in Brussels)
Has 25 members and
2 observers
Governed by the PRACE
Council in which each
member has a seat
Daily management
of the association is
delegated to the Board
of Directors
[26] PRACE
Mission:
enabling world-class science through
large scale simulations
Offering:
HPC resources on leading edge
capability systems
Resource award:
through a single and fair pan-
European peer review process for
open research
HPC
[26] PRACE
Selected Facts for more pieces of Information about HPC Techniques & Methods
More than 10 000 people trained by 6 PRACE Advanced Training Centers (PATC) and other events
Training portal consists of valuable material in all fields related to HPC & supercomputing
Easy search
function to
find materials
of past events
Material of this
training will be
also available
after the event
© Photograph by Rama,
Wikimedia Commons
1.000.000.000.000.000 FLOP/s
~295.000 cores~2009 (JUGENE)
>5.900.000.000.000.000
FLOP/s
~ 500.000 cores
~2013 end of service in 2018
Clock-rate for single processors increased from 10 MHz (Intel 286) to 4 GHz (Pentium 4) in 30 years
Clock rate increase with higher 5 GHz unfortunately reached a limit due to power limitations / heat
Multi-core CPU chips have quad, six, or n processing cores on one chip and use cache hierarchies
Strong collaboration
with our industry partners
Intel, Extoll & Megware
27 partners
Coordinated by JSC
Juelich Supercomputing Centre
EU-funding: 30 M€ implements the DEEP projects
designs in its HPC infrastructure
JSC-part > 5,3 M€
Nov 2011 – Dec 2020
Backup: In the case we use a SSH client (MobaXterm) Navigate to your practical folder
Open the terminal cd /p/project/training2001/$USER/practicals
Start an interactive session (Note the reservation: 17/02
Create a new folder where you can store ‘prace_1_gpu’, 18/02 ‘prace_2_gpu’, 19/02 ‘prace_3_gpu’)
all your personal items (practicals and new files) salloc --gres=gpu:1 --partition=gpus --nodes=1 --
mkdir /p/project/training2001/$USER account=training2001 --time=01:00:00 --reservation=
prace_1_gpu
Copy your own copy of the practicals Run the pre-made script which will load the necessary modules
cd /p/project/training2001/$USER and activate the python virtual enviroment
cp -R /p/project/training2001/practicals . . run_venv_jupyter_terminal.sh
Done, now you can run your python scripts by using
srun python name_function.py
[1] Morris Riedel, ‘Deep Learning - Using a Convolutional Neural Network‘, Invited YouTube Lecture, six lectures & exercises, University of Ghent, 2017, Online:
https://www.youtube.com/watch?v=gOL1_YIosYk&list=PLrmNhuZo9sgZUdaZ-f6OHK2yFW1kTS2qF
[2] M. Riedel et al., ‘Introduction to Deep Learning Models‘, JSC Tutorial, three days, JSC, 2019, Online:
http://www.morrisriedel.de/introduction-to-deep-learning-models
[3] H. Lee et al., ‘Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations’, Online:
http://doi.acm.org/10.1145/1553374.1553453
[4] YouTube Video, ‘Neural Network 3D Simulation‘, Online:
https://www.youtube.com/watch?v=3JQ3hYko51Y
[5] A. Rosebrock, ‘Get off the deep learning bandwagon and get some perspective‘, Online:
http://www.pyimagesearch.com/2014/06/09/get-deep-learning-bandwagon-get-perspective/
[6] Big Data Tips – Big Data Mining & Machine Learning, Online:
http://www.big-data.tips/
[7] NVIDIA Web Page, Online:
https://www.nvidia.com/en-us/
[8] Keras Python High-Level Deep Learning Library, Online:
https://keras.io/
[9] TensorFlow Python Low-Level Deep learning Library, Online:
https://www.tensorflow.org/
[10] Deep Learning Start-Up Beispiel Deutschland, Online:
https://soccerwatch.tv/
[11] C. Bodenstein, M. Goetz, M. Riedel, ‘Automated Soccer Scene Tracking using Deep Neural Networks’, Poster IAS Symposium, Online:
https://www.researchgate.net/publication/328997974_Automated_Soccer_Scene_Tracking_Using_Deep_Neural_Networks
PD Dr. Senior PhD Senior PhD PhD Student PhD Student PhD Student
G. Cavallaro Student A.S. Memon Student M.S. Memon E. Erlingsson S. Bakarat R. Sedona
DEEP
Finished PHD Thesis Thesis Learning
in 2018 Completed Completed Startup