SlideShare a Scribd company logo
Optimize your Machine Learning workloads  | AWS Summit Tel Aviv 2019
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Optimize your Machine Learning workloads
Julien Simon
Global Evangelist, AI & Machine Learning
@julsimon
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Put machine learning in the
hands of every developer
Our mission at AWS
Now let’s make it as
fast, efficient and unexpensive
as possible
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Amazon EC2 P3dn
https://aws.amazon.com/blogs/aws/new-ec2-p3dn-gpu-instances-with-100-gbps-networking-local-nvme-storage-for-faster-machine-learning-p3-price-reduction/
Reduce machine learning
training time
Better GPU
utilization
Support larger, more
complex models
K E Y F E AT U R E S
100Gbps of networking
bandwidth
8 NVIDIA Tesla V100
GPUs
32GB of
memory per GPU
(2x more P3)
96 Intel
Skylake vCPUs
(50% more than P3)
with AVX-512
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Amazon EC2 C5n
https://aws.amazon.com/blogs/aws/new-c5n-instances-with-100-gbps-networking/
Intel Xeon Platinum 8000
Up to 3.5GHz single core speed
Up to 100Gbit networking
Based on Nitro hypervisor for
bare metal-like performance
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Making TensorFlow faster
Training a ResNet-50 benchmark with the synthetic ImageNet dataset using
our optimized build of TensorFlow 1.11 on a c5.18xlarge instance type is 11x
faster than training on the stock binaries.
https://aws.amazon.com/about-aws/whats-new/2018/10/chainer4-4_theano_1-0-2_launch_deep_learning_ami/
October 2018
Available with Amazon SageMaker and the
AWS Deep Learning AMIs
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Scaling TensorFlow near-linearly to 256 GPUs
https://aws.amazon.com/about-aws/whats-new/2018/11/tensorflow-scalability-to-256-gpus/
Stock
TensorFlow
65%
scaling efficiency
with 256 GPUs
30m
training time
AWS-Optimized
TensorFlow
90%
scaling efficiency
with 256 GPUs
Available with Amazon
SageMaker and the
AWS Deep Learning
AMIs
14m
training time
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Dynamic training with Apache MXNet
https://aws.amazon.com/blogs/machine-learning/introducing-dynamic-training-for-deep-learning-with-amazon-ec2/
Use a variable number of instances
for distributed training
No loss of accuracy
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Examples of hyperparameters
Neural Networks
Number of layers
Hidden layer width
Learning rate
Embedding dimensions
Dropout
…
Decision Trees
Tree depth
Max leaf nodes
Gamma
Eta
Lambda
Alpha
…
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Automatic Model Tuning
Finding the optimal set of hyper parameters
1. Manual Search (”I know what I’m doing”)
2. Grid Search (“X marks the spot”)
Typically training hundreds of models
Slow and expensive
3. Random Search (“Spray and pray”)
Works better and faster than Grid Search
But… but… but… it’s random!
4. HPO: use Machine Learning
• Training fewer models
• Gaussian Process Regression and Bayesian Optimization
• You can now resume from a previous tuning job
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Hardware optimization is extremely complex
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Amazon Neo: compiling models
https://aws.amazon.com/blogs/aws/amazon-sagemaker-neo-train-your-machine-learning-models-once-run-them-anywhere/
• Train once, run anywhere
• Frameworks and algorithms
• TensorFlow, Apache MXNet, PyTorch, ONNX, and XGBoost
• Hardware architectures
• ARM, Intel, and NVIDIA starting today
• Cadence, Qualcomm, and Xilinx hardware coming soon
• Amazon SageMaker Neo is open source, enabling hardware vendors to
customize it for their processors and devices: https://github.com/neo-ai
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Demo:
Compiling ResNet-50 for the Raspberry Pi
Configure the compilation job
{
"RoleArn":$ROLE_ARN,
"InputConfig": {
"S3Uri":"s3://jsimon-neo/model.tar.gz",
"DataInputConfig": "{"data": [1, 3, 224, 224]}",
"Framework": "MXNET"
},
"OutputConfig": {
"S3OutputLocation": "s3://jsimon-neo/",
"TargetDevice": "rasp3b"
},
"StoppingCondition": {
"MaxRuntimeInSeconds": 300
}
}
Compile the model
$ aws sagemaker create-compilation-job
--cli-input-json file://config.json
--compilation-job-name resnet50-mxnet-pi
$ aws s3 cp s3://jsimon-neo/model-
rasp3b.tar.gz .
$ gtar tfz model-rasp3b.tar.gz
compiled.params
compiled_model.json
compiled.so
Predict with the compiled model
from dlr import DLRModel
model = DLRModel('resnet50', input_shape,
output_shape, device)
out = model.run(input_data)
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Compressing deep learning models
• Compression is the process of reducing the size of a trained network, either by
removing certain layers or by shrinking layers, while maintaining accuracy.
• A smaller model will predict faster and require less memory.
• The number of possible combinations makes is difficult to perform this task
manually, or even programmatically.
• Reinforcement learning to the rescue!
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Defining the problem
• Objective: find the smallest possible network
architecture from a pre-trained network
architecture, while producing the best
accuracy.
• Environment: a custom developed environment
that accepts a Boolean array of layers to
remove from the RL agent and produces an
observation describing layers.
• State: the layers.
• Action: A boolean array one for each layer.
• Reward: a combination of compression ratio
and accuracy.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
https://github.com/awslabs/amazon-sagemaker-
examples/tree/master/reinforcement_learning/rl_network_compression_ray_custom
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Inference
90%
Training
10%
Predictions drive
complexity and
cost in production
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Are you making the most of your infrastructure?
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Amazon Elastic Inference
https://aws.amazon.com/blogs/aws/amazon-elastic-inference-gpu-powered-deep-learning-inference-acceleration/
Match capacity
to demand
Available between 1 to 32
TFLOPS
K E Y F E AT U R E S
Integrated with
Amazon EC2,
Amazon SageMaker, and
Amazon DL AMIs
Support for TensorFlow, Apache
MXNet, and ONNX
with PyTorch coming soon
Single and
mixed-precision
operations
Lower inference costs up
to 75%
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Faster training
Faster inference
Save time and money
No plumbing
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Getting started
https://ml.aws
https://aws.amazon.com/sagemaker
https://github.com/awslabs/amazon-sagemaker-examples
https://medium.com/@julsimon
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Thank you!
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Julien Simon
Global Evangelist, AI & Machine Learning
@julsimon

More Related Content

Similar to Optimize your Machine Learning workloads | AWS Summit Tel Aviv 2019 (20)

PPTX
Optimize your machine learning workloads on AWS (March 2019)
Julien SIMON
 
PPTX
Optimize your Machine Learning workloads (April 2019)
Julien SIMON
 
PPTX
Optimize your Machine Learning Workloads on AWS (July 2019)
Julien SIMON
 
PPTX
Build, train, and deploy Machine Learning models at scale (May 2018)
Julien SIMON
 
PPTX
Amazon SageMaker (December 2018)
Julien SIMON
 
PDF
Julien Simon, Principal Technical Evangelist at Amazon - Machine Learning: Fr...
Codiax
 
PPTX
An Introduction to Amazon SageMaker (October 2018)
Julien SIMON
 
PPTX
Build, train, and deploy Machine Learning models at scale (May 2018)
Julien SIMON
 
PPTX
AWS re:Invent 2018 - Machine Learning recap (December 2018)
Julien SIMON
 
PPTX
Speeding up Deep Learning training and inference
Thomas Delteil
 
PPTX
AWS re:Invent 2018 - AIM401 - Deep Learning using Tensorflow
Julien SIMON
 
PPTX
AWS re:Invent 2018 - ENT321 - SageMaker Workshop
Julien SIMON
 
PDF
Deep Learning with Tensorflow and Apache MXNet on AWS (April 2019)
Julien SIMON
 
PPTX
Deep Learning with TensorFlow and Apache MXNet on Amazon SageMaker (March 2019)
Julien SIMON
 
PPTX
Build, train and deploy ML models with SageMaker (October 2019)
Julien SIMON
 
PDF
DataXDay - Machine learning models at scale with Amazon SageMaker
DataXDay Conference by Xebia
 
PPTX
Advanced Machine Learning with Amazon SageMaker
Julien SIMON
 
PPTX
Deep learning acceleration with Amazon Elastic Inference
Hagay Lupesko
 
PDF
Deep Learning 모델의 효과적인 분산 트레이닝과 모델 최적화 방법 - 김무현 데이터 사이언티스트, AWS :: AWS Summit...
Amazon Web Services Korea
 
PPTX
Quickly and easily build, train, and deploy machine learning models at any scale
AWS Germany
 
Optimize your machine learning workloads on AWS (March 2019)
Julien SIMON
 
Optimize your Machine Learning workloads (April 2019)
Julien SIMON
 
Optimize your Machine Learning Workloads on AWS (July 2019)
Julien SIMON
 
Build, train, and deploy Machine Learning models at scale (May 2018)
Julien SIMON
 
Amazon SageMaker (December 2018)
Julien SIMON
 
Julien Simon, Principal Technical Evangelist at Amazon - Machine Learning: Fr...
Codiax
 
An Introduction to Amazon SageMaker (October 2018)
Julien SIMON
 
Build, train, and deploy Machine Learning models at scale (May 2018)
Julien SIMON
 
AWS re:Invent 2018 - Machine Learning recap (December 2018)
Julien SIMON
 
Speeding up Deep Learning training and inference
Thomas Delteil
 
AWS re:Invent 2018 - AIM401 - Deep Learning using Tensorflow
Julien SIMON
 
AWS re:Invent 2018 - ENT321 - SageMaker Workshop
Julien SIMON
 
Deep Learning with Tensorflow and Apache MXNet on AWS (April 2019)
Julien SIMON
 
Deep Learning with TensorFlow and Apache MXNet on Amazon SageMaker (March 2019)
Julien SIMON
 
Build, train and deploy ML models with SageMaker (October 2019)
Julien SIMON
 
DataXDay - Machine learning models at scale with Amazon SageMaker
DataXDay Conference by Xebia
 
Advanced Machine Learning with Amazon SageMaker
Julien SIMON
 
Deep learning acceleration with Amazon Elastic Inference
Hagay Lupesko
 
Deep Learning 모델의 효과적인 분산 트레이닝과 모델 최적화 방법 - 김무현 데이터 사이언티스트, AWS :: AWS Summit...
Amazon Web Services Korea
 
Quickly and easily build, train, and deploy machine learning models at any scale
AWS Germany
 

More from AWS Summits (20)

PDF
AWS Summit Singapore 2019 | The Smart Way to Build an AI & ML Strategy for Yo...
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Bridging Start-ups and Enterprises
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Hiring a Global Rock Star Team: Tips and Tricks
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Five Common Technical Challenges for Startups
AWS Summits
 
PDF
AWS Summit Singapore 2019 | A Founder's Journey to Exit
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Realising Business Value with AWS Analytics Services
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Snowflake: Your Data. No Limits
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Amazon Digital User Engagement Solutions
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Driving Business Outcomes with Data Lake on AWS
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Big Data Analytics Architectural Patterns and Bes...
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Microsoft DevOps on AWS
AWS Summits
 
PDF
AWS Summit Singapore 2019 | The Serverless Lifecycle: Development and Operati...
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Accelerating Enterprise Cloud Transformation by M...
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Operating Microservices at Hyperscale
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Autoscaling Your Kubernetes Workloads
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Realising Business Value
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Latest Trends for Cloud-Native Application Develo...
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Transformation Towards a Digital Native Enterprise
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Pragmatic Container Security
AWS Summits
 
PDF
AWS Summit Singapore 2019 | Enterprise Migration Journey Roadmap
AWS Summits
 
AWS Summit Singapore 2019 | The Smart Way to Build an AI & ML Strategy for Yo...
AWS Summits
 
AWS Summit Singapore 2019 | Bridging Start-ups and Enterprises
AWS Summits
 
AWS Summit Singapore 2019 | Hiring a Global Rock Star Team: Tips and Tricks
AWS Summits
 
AWS Summit Singapore 2019 | Five Common Technical Challenges for Startups
AWS Summits
 
AWS Summit Singapore 2019 | A Founder's Journey to Exit
AWS Summits
 
AWS Summit Singapore 2019 | Realising Business Value with AWS Analytics Services
AWS Summits
 
AWS Summit Singapore 2019 | Snowflake: Your Data. No Limits
AWS Summits
 
AWS Summit Singapore 2019 | Amazon Digital User Engagement Solutions
AWS Summits
 
AWS Summit Singapore 2019 | Driving Business Outcomes with Data Lake on AWS
AWS Summits
 
AWS Summit Singapore 2019 | Big Data Analytics Architectural Patterns and Bes...
AWS Summits
 
AWS Summit Singapore 2019 | Microsoft DevOps on AWS
AWS Summits
 
AWS Summit Singapore 2019 | The Serverless Lifecycle: Development and Operati...
AWS Summits
 
AWS Summit Singapore 2019 | Accelerating Enterprise Cloud Transformation by M...
AWS Summits
 
AWS Summit Singapore 2019 | Operating Microservices at Hyperscale
AWS Summits
 
AWS Summit Singapore 2019 | Autoscaling Your Kubernetes Workloads
AWS Summits
 
AWS Summit Singapore 2019 | Realising Business Value
AWS Summits
 
AWS Summit Singapore 2019 | Latest Trends for Cloud-Native Application Develo...
AWS Summits
 
AWS Summit Singapore 2019 | Transformation Towards a Digital Native Enterprise
AWS Summits
 
AWS Summit Singapore 2019 | Pragmatic Container Security
AWS Summits
 
AWS Summit Singapore 2019 | Enterprise Migration Journey Roadmap
AWS Summits
 
Ad

Optimize your Machine Learning workloads | AWS Summit Tel Aviv 2019

  • 2. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Optimize your Machine Learning workloads Julien Simon Global Evangelist, AI & Machine Learning @julsimon
  • 3. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Put machine learning in the hands of every developer Our mission at AWS Now let’s make it as fast, efficient and unexpensive as possible
  • 4. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 5. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Amazon EC2 P3dn https://aws.amazon.com/blogs/aws/new-ec2-p3dn-gpu-instances-with-100-gbps-networking-local-nvme-storage-for-faster-machine-learning-p3-price-reduction/ Reduce machine learning training time Better GPU utilization Support larger, more complex models K E Y F E AT U R E S 100Gbps of networking bandwidth 8 NVIDIA Tesla V100 GPUs 32GB of memory per GPU (2x more P3) 96 Intel Skylake vCPUs (50% more than P3) with AVX-512
  • 6. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Amazon EC2 C5n https://aws.amazon.com/blogs/aws/new-c5n-instances-with-100-gbps-networking/ Intel Xeon Platinum 8000 Up to 3.5GHz single core speed Up to 100Gbit networking Based on Nitro hypervisor for bare metal-like performance
  • 7. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Making TensorFlow faster Training a ResNet-50 benchmark with the synthetic ImageNet dataset using our optimized build of TensorFlow 1.11 on a c5.18xlarge instance type is 11x faster than training on the stock binaries. https://aws.amazon.com/about-aws/whats-new/2018/10/chainer4-4_theano_1-0-2_launch_deep_learning_ami/ October 2018 Available with Amazon SageMaker and the AWS Deep Learning AMIs
  • 8. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Scaling TensorFlow near-linearly to 256 GPUs https://aws.amazon.com/about-aws/whats-new/2018/11/tensorflow-scalability-to-256-gpus/ Stock TensorFlow 65% scaling efficiency with 256 GPUs 30m training time AWS-Optimized TensorFlow 90% scaling efficiency with 256 GPUs Available with Amazon SageMaker and the AWS Deep Learning AMIs 14m training time
  • 9. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Dynamic training with Apache MXNet https://aws.amazon.com/blogs/machine-learning/introducing-dynamic-training-for-deep-learning-with-amazon-ec2/ Use a variable number of instances for distributed training No loss of accuracy
  • 10. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 11. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Examples of hyperparameters Neural Networks Number of layers Hidden layer width Learning rate Embedding dimensions Dropout … Decision Trees Tree depth Max leaf nodes Gamma Eta Lambda Alpha …
  • 12. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Automatic Model Tuning Finding the optimal set of hyper parameters 1. Manual Search (”I know what I’m doing”) 2. Grid Search (“X marks the spot”) Typically training hundreds of models Slow and expensive 3. Random Search (“Spray and pray”) Works better and faster than Grid Search But… but… but… it’s random! 4. HPO: use Machine Learning • Training fewer models • Gaussian Process Regression and Bayesian Optimization • You can now resume from a previous tuning job
  • 13. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 14. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Hardware optimization is extremely complex
  • 15. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Amazon Neo: compiling models https://aws.amazon.com/blogs/aws/amazon-sagemaker-neo-train-your-machine-learning-models-once-run-them-anywhere/ • Train once, run anywhere • Frameworks and algorithms • TensorFlow, Apache MXNet, PyTorch, ONNX, and XGBoost • Hardware architectures • ARM, Intel, and NVIDIA starting today • Cadence, Qualcomm, and Xilinx hardware coming soon • Amazon SageMaker Neo is open source, enabling hardware vendors to customize it for their processors and devices: https://github.com/neo-ai
  • 16. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Demo: Compiling ResNet-50 for the Raspberry Pi Configure the compilation job { "RoleArn":$ROLE_ARN, "InputConfig": { "S3Uri":"s3://jsimon-neo/model.tar.gz", "DataInputConfig": "{"data": [1, 3, 224, 224]}", "Framework": "MXNET" }, "OutputConfig": { "S3OutputLocation": "s3://jsimon-neo/", "TargetDevice": "rasp3b" }, "StoppingCondition": { "MaxRuntimeInSeconds": 300 } } Compile the model $ aws sagemaker create-compilation-job --cli-input-json file://config.json --compilation-job-name resnet50-mxnet-pi $ aws s3 cp s3://jsimon-neo/model- rasp3b.tar.gz . $ gtar tfz model-rasp3b.tar.gz compiled.params compiled_model.json compiled.so Predict with the compiled model from dlr import DLRModel model = DLRModel('resnet50', input_shape, output_shape, device) out = model.run(input_data)
  • 17. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 18. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Compressing deep learning models • Compression is the process of reducing the size of a trained network, either by removing certain layers or by shrinking layers, while maintaining accuracy. • A smaller model will predict faster and require less memory. • The number of possible combinations makes is difficult to perform this task manually, or even programmatically. • Reinforcement learning to the rescue!
  • 19. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Defining the problem • Objective: find the smallest possible network architecture from a pre-trained network architecture, while producing the best accuracy. • Environment: a custom developed environment that accepts a Boolean array of layers to remove from the RL agent and produces an observation describing layers. • State: the layers. • Action: A boolean array one for each layer. • Reward: a combination of compression ratio and accuracy.
  • 20. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. https://github.com/awslabs/amazon-sagemaker- examples/tree/master/reinforcement_learning/rl_network_compression_ray_custom
  • 21. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 22. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Inference 90% Training 10% Predictions drive complexity and cost in production
  • 23. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Are you making the most of your infrastructure?
  • 24. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Amazon Elastic Inference https://aws.amazon.com/blogs/aws/amazon-elastic-inference-gpu-powered-deep-learning-inference-acceleration/ Match capacity to demand Available between 1 to 32 TFLOPS K E Y F E AT U R E S Integrated with Amazon EC2, Amazon SageMaker, and Amazon DL AMIs Support for TensorFlow, Apache MXNet, and ONNX with PyTorch coming soon Single and mixed-precision operations Lower inference costs up to 75%
  • 25. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 26. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Faster training Faster inference Save time and money No plumbing
  • 27. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Getting started https://ml.aws https://aws.amazon.com/sagemaker https://github.com/awslabs/amazon-sagemaker-examples https://medium.com/@julsimon
  • 28. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 29. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Thank you! © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Julien Simon Global Evangelist, AI & Machine Learning @julsimon