The document discusses OpenVINOTM, an Intel toolkit that provides high performance computer vision and deep learning inference capabilities. It allows building applications that leverage OpenCV, deep learning models, and heterogeneous execution across CPU, GPU, FPGA and VPU hardware. Key benefits include portable deployment across platforms with a minimal footprint, optimized performance on Intel hardware, and pre-trained models for common tasks like object detection. The toolkit includes libraries, tools for model optimization, and samples to help developers build and deploy high performance computer vision and deep learning applications.
The document discusses Intel's OpenVINOTM toolkit, which provides tools to optimize deep learning models for deployment across Intel platforms. It notes that the toolkit has been adopted by over 30 customers and provides up to 5x faster and more efficient inference compared to frameworks like TensorFlow. The toolkit allows models to run on CPU, GPU, VPU and FPGA with minimal code changes. It also supports post-training quantization, video preprocessing and asynchronous execution for improved performance. The document encourages selecting models suited for deployment and using the included Open Model Zoo for pre-trained models optimized for Intel hardware.
The dominant sequence transduction models are based on complex recurrent or convolutional neural networks in an encoder-decoder configuration. The best performing models also connect the encoder and decoder through an attention mechanism. We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two machine translation tasks show these models to be superior in quality while being more parallelizable and requiring significantly less time to train.
Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task, improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new single-model state-of-the-art BLEU score of 41.0 after training for 3.5 days on eight GPUs, a small fraction of the training costs of the best models from the literature. We show that the Transformer generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data.
The content was modified from Google Content Group
Eric ShangKuan([email protected])
---
TensorFlow Lite guide( for mobile & IoT )
TensorFlow Lite is a set of tools to help developers run TensorFlow models on mobile, embedded, and IoT devices. It enables on-device machine learning inference with low latency and small binary size.
TensorFlow Lite consists of two main components:
The TensorFlow Lite interpreter:
- optimize models on many different hardware types, like mobile phones, embedded Linux devices, and microcontrollers.
The TensorFlow Lite converter:
- which converts TensorFlow models into an efficient form for use by the interpreter, and can introduce optimizations to improve binary size and performance.
---
Event: PyLadies TensorFlow All-Around
Date: Sep 25, 2019
Event link: https://www.meetup.com/PyLadies-Berlin/events/264205538/
Linkedin: http://linkedin.com/in/mia-chang/
Machine learning can be distributed across multiple machines to allow for processing of large datasets and complex models. There are three main approaches to distributed machine learning: data parallel, where the data is partitioned across machines and models are replicated; model parallel, where different parts of large models are distributed; and graph parallel, where graphs and algorithms are partitioned. Distributed frameworks use these approaches to efficiently and scalably train machine learning models on big data in parallel.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2022/06/optimization-techniques-with-intels-openvino-to-enhance-performance-on-your-existing-hardware-a-presentation-from-intel/
Nico Galoppo, Principal Engineer (substituting for Ansley Dunn, Product Marketing Manager), and Ryan Loney, Technical Product Manager, both of Intel, present the “Optimization Techniques with Intel’s OpenVINO to Enhance Performance on Your Existing Hardware” tutorial at the May 2022 Embedded Vision Summit.
Whether you’re using TensorFlow, PyTorch or another framework, Galoppo and Loney show you optimization techniques to enhance performance on your existing hardware. With the OpenVINO Toolkit, built on the foundation of OneAPI, developers can utilize their own AI model or leverage one of the hundreds of pre-trained models available across vision and audio use cases.
In this presentation, you’ll learn how the Neural Network Compression Framework provides optimal model training templates for performance boosts while preserving accuracy, and how the Model Optimizer reduces complexity and makes model conversion faster. Other areas explored by Galoppo and Loney include auto device discovery to enable automatic load balancing and how to optimize for latency or throughput based on your workload.
The document introduces artificial intelligence, machine learning, and deep learning. It discusses supervised, unsupervised, and reinforced learning techniques. Examples of applications discussed include image recognition, natural language processing, and virtual assistants. The document also notes that some AI systems have developed their own internal languages when interacting without human supervision.
CUDA is a parallel computing platform and programming model developed by Nvidia that allows software developers and researchers to utilize GPUs for general purpose processing. CUDA allows developers to achieve up to 100x performance gains over CPU-only applications. CUDA works by having the CPU copy input data to GPU memory, executing a kernel program on the GPU that runs in parallel across many threads, and copying the results back to CPU memory. Key GPU memories that can be used in CUDA programs include shared memory for thread cooperation, textures for cached reads, and constants for read-only data.
Distributed Caching in Kubernetes with HazelcastMesut Celik
As Monolith to Microservices migration almost became mainstream, Engineering Teams have to think about how their caching strategies will evolve in cloud-native world. Kubernetes is clear winner in containerized world so caching solutions must be cloud-ready and natural fit for Kubernetes.
Caching is an important piece in high performance microservices and choosing right architectural pattern can be crucial for your deployments. Hazelcast is a well known caching solution in open source community and can handle caching piece in microservices based applications.
In this talk, you will learn
* Distributed Caching With Hazelcast
* Distributed Caching Patterns in Kubernetes
* Kubernetes Deployment Options and Best Practices
* How to Handle Distributed Caching Day 2 Operations
Wix' internal ML Platform, whose mission is to allow data scientists and analysts at Wix to build, deploy, maintain, and monitor machine learning models in production with minimal engineering efforts
TensorRT is an NVIDIA tool that optimizes and accelerates deep learning models for production deployment. It performs optimizations like layer fusion, reduced precision from FP32 to FP16 and INT8, kernel auto-tuning, and multi-stream execution. These optimizations reduce latency and increase throughput. TensorRT automatically optimizes models by taking in a graph, performing optimizations, and outputting an optimized runtime engine.
RESTful Machine Learning with Flask and TensorFlow Serving - Carlo MazzaferroPyData
Those of us who use TensorFlow often focus on building the model that's most predictive, not the one that's most deployable. So how to put that hard work to work? In this talk, we'll walk through a strategy for taking your machine learning models from Jupyter Notebook into production and beyond.
What is Deep Learning | Deep Learning Simplified | Deep Learning Tutorial | E...Edureka!
This Edureka "What is Deep Learning" video will help you to understand about the relationship between Deep Learning, Machine Learning and Artificial Intelligence and how Deep Learning came into the picture. This tutorial will be discussing about Artificial Intelligence, Machine Learning and its limitations, how Deep Learning overcame Machine Learning limitations and different real-life applications of Deep Learning.
Below are the topics covered in this tutorial:
1. What Is Artificial Intelligence?
2. What Is Machine Learning?
3. Limitations Of Machine Learning
4. Deep Learning To The Rescue
5. What Is Deep Learning?
6. Deep Learning Applications
To take a structured training on Deep Learning, you can check complete details of our Deep Learning with TensorFlow course here: https://goo.gl/VeYiQZ
An introduction to the Transformers architecture and BERTSuman Debnath
The transformer is one of the most popular state-of-the-art deep (SOTA) learning architectures that is mostly used for natural language processing (NLP) tasks. Ever since the advent of the transformer, it has replaced RNN and LSTM for various tasks. The transformer also created a major breakthrough in the field of NLP and also paved the way for new revolutionary architectures such as BERT.
OpenGL is a powerful, low-level graphics toolkit with a steep learning curve that allows access to accelerated GPU hardware. Using OpenGL, developers achieve high fidelity, animated graphics ubiquitous in games, screen productions and scientific software.
Join us for a one-hour webinar and we will give a comprehensive overview of the many aspects of OpenGL development where Qt provides advanced interfaces that let the developer focus on the tasks instead of dealing with repetitive and error-prone, platform dependent issues.
This document provides an overview of ONNX and ONNX Runtime. ONNX is an open format for machine learning models that allows models to be shared across different frameworks and tools. ONNX Runtime is a cross-platform open source inference engine that runs ONNX models. It supports hardware acceleration and has a modular design that allows for custom operators and execution providers to extend its capabilities. The document discusses how ONNX helps with deploying machine learning models from research to production and how ONNX Runtime performs high performance inference through optimizations and hardware acceleration.
NVIDIA’s invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI — the next era of computing — with the GPU acting as the brain of computers, robots, and self-driving cars that can perceive and understand the world. Today, NVIDIA is increasingly known as “the AI computing company.”
Deep learning is a branch of machine learning that uses neural networks with multiple processing layers to learn representations of data with multiple levels of abstraction. It has been applied to problems like image recognition, natural language processing, and game playing. Deep learning architectures like deep neural networks use techniques like pretraining, dropout, and early stopping to avoid overfitting. Popular deep learning frameworks and libraries include TensorFlow, Keras, and PyTorch.
The RAPIDS suite of software libraries gives you the freedom to execute end-to-end data science and analytics pipelines entirely on GPUs. It relies on NVIDIA® CUDA® primitives for low-level compute optimization, but exposes that GPU parallelism and high-bandwidth memory speed through user-friendly Python interfaces.
Machine Learning. What is machine learning. Normal computer vs ML. Types of Machine Learning. Some ML Object detection methods. Faster CNN, RCNN, YOLO, SSD. Real Life ML Applications. Best Programming Languages for ML. Difference Between Machine Learning And Artificial Intelligence. Advantages of Machine Learning. Disadvantages of Machine Learning
While Go is the language-of-choice in the cloud-native world, Python has a huge community and makes it really easy to extend Kubernetes in only a few lines of code.
This talk shows examples on how to use Python to query the Kubernetes API, how to write simple controllers in only 10 lines of Python, how to build complete web UIs, and how to test everything with py.test and Kind.
Some of the open-source projects which will be covered: pykube-ng, Kubernetes Web View, kube-janitor, and Kopf (Kubernetes Operator Pythonic Framework).
Talk held in Prague on 2019-09-05:
https://www.meetup.com/Cloud-Native-Prague/events/263802447/
Slides reviewing the paper:
Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, and Illia Polosukhin. "Attention is all you need." In Advances in Neural Information Processing Systems, pp. 6000-6010. 2017.
The dominant sequence transduction models are based on complex recurrent orconvolutional neural networks in an encoder and decoder configuration. The best performing such models also connect the encoder and decoder through an attentionm echanisms. We propose a novel, simple network architecture based solely onan attention mechanism, dispensing with recurrence and convolutions entirely.Experiments on two machine translation tasks show these models to be superiorin quality while being more parallelizable and requiring significantly less timeto train. Our single model with 165 million parameters, achieves 27.5 BLEU onEnglish-to-German translation, improving over the existing best ensemble result by over 1 BLEU. On English-to-French translation, we outperform the previoussingle state-of-the-art with model by 0.7 BLEU, achieving a BLEU score of 41.1.
This document provides an introduction to Qt Quick, the declarative UI framework in Qt. It discusses key Qt Quick concepts like QML syntax, properties, binding, and user interface composition. It also covers cross-platform development with Qt, the Qt module system, and Qt UI technologies like Qt Quick, Qt Widgets and web/hybrid approaches. Finally, it discusses specific Qt Quick topics like elements, properties, identities, anchors, and user interaction handling.
Using MLOps to Bring ML to Production/The Promise of MLOpsWeaveworks
In this final Weave Online User Group of 2019, David Aronchick asks: have you ever struggled with having different environments to build, train and serve ML models, and how to orchestrate between them? While DevOps and GitOps have made huge traction in recent years, many customers struggle to apply these practices to ML workloads. This talk will focus on the ways MLOps has helped to effectively infuse AI into production-grade applications through establishing practices around model reproducibility, validation, versioning/tracking, and safe/compliant deployment. We will also talk about the direction for MLOps as an industry, and how we can use it to move faster, with more stability, than ever before.
The recording of this session is on our YouTube Channel here: https://youtu.be/twsxcwgB0ZQ
Speaker: David Aronchick, Head of Open Source ML Strategy, Microsoft
Bio: David leads Open Source Machine Learning Strategy at Azure. This means he spends most of his time helping humans to convince machines to be smarter. He is only moderately successful at this. Previously, David led product management for Kubernetes at Google, launched GKE, and co-founded the Kubeflow project. David has also worked at Microsoft, Amazon and Chef and co-founded three startups.
Sign up for a free Machine Learning Ops Workshop: http://bit.ly/MLOps_Workshop_List
Weaveworks will cover concepts such as GitOps (operations by pull request), Progressive Delivery (canary, A/B, blue-green), and how to apply those approaches to your machine learning operations to mitigate risk.
The objectives of the seminar are to shed a light on the premises of FP and give you a basic understanding of the pillars of FP so that you would feel enlightened at the end of the session. When you walk away from the seminar you should feel an inner light about the new way of programming and an urge & motivation to code like you never before did!
Functional programming should not be confused with imperative (or procedural) programming. Neither it is like object oriented programming. It is something different. Not radically so, since the concepts that we will be exploring are familiar programming concepts, just expressed in a different way. The philosophy behind how these concepts are applied to solving problems are also a little different. We shall learn and talk about essentially the fundamental elements of Functional Programming.
Slide 1
TypeScript
* This presentation is to show TypeScript's major feature and the benefit that it brings to your JavaScript projects.
* Our main objective is just to spark interest especially to those not familiar with the tool.
Slide 2
- What is TypeScript
* go to next slide
Slide 3
- Is a superset of JavaScript
* it simply means an extension to JavaScript
- Use JavaScript code on TypeScript
* JS code naturally works on TypeScript
* Which also means your beloved JavaScript libraries such as JQuery, or your fancy interacive plugins would work as well.
- TypeScript compiles to plain old JavaScript
* TS code compiles to simple and clean JS.
Slide 4
- Screenshot of TS compiled to JS
* In this example, compiling a TS class code would result to a JS version, and a regular JavaScript function when compiled is basically untouched.
Slide 5
- TypeScript's Main Feature
* So what does TS provide us with? What does it actually do?
Slide 6
- Static Type Checking
* TypeScript allows us to enable type checking by defining data types to your for ex. variables, function parameters and return types.
Slide 7
- Screenshot of basic Static Type Checking
* In this example…
* What I've done here was to assign supposedly wrong values for what the variables or parameters were meant to hold
* As JavaScript is a dynamic and untyped language these expressions would either fail or be okay when you run it on your browser.
* In TypeScript by enabling static type checking these potential errors are caught earlier (see the red marks on the expressions) and wouldn't even allow you to compile unless these are resolved.
* In addition you can also type arrays and object literals
Slide 8
- Effects of Static Type Checking
* As TS code is statically type-checked a side effect of such...
- Allows IDEs to perform live error checks
- Exposes auto-completion and code hinting
Slide 9
- Screenshot of code hinting
* Say I was coding JQuery on regular JavaScript code there would be no natural way to help me identify its class properties, methods and parameters... except through reading the API documentation or a separate plugin.
* As a result of static type checking this allows IDE's to access these class members as code hints
* So if this was a 3rd party library how much more if you are just referencing your own JavaScript/TypeScript files within your project.
Slide 10
- A few of the other cool features
* That was only the basic feature of TypeScript
* A few of the other cool features are...
Slide 11
- End
How to Get the Best Deep Learning performance with OpenVINO ToolkitYury Gorbachev
The document discusses Intel's OpenVINOTM toolkit, which provides tools to optimize deep learning models for deployment across Intel platforms. It notes that the toolkit has been adopted by over 30 customers and provides up to 5x faster and more efficient inference compared to frameworks like TensorFlow and MXNet. The toolkit allows models to run on CPU, GPU, VPU and FPGA with minimal code changes and supports post-training quantization and binarization. It includes tools for video preprocessing and benchmarks to measure application performance rather than just inference speed. The document recommends using the included high performance models from the Open Model Zoo and demo applications.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/itseez/embedded-vision-training/videos/pages/may-2015-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Yury Gorbachev, Principal Engineer at itseez, presents the "OpenCV for Embedded: Lessons Learned" tutorial at the May 2015 Embedded Vision Summit.
OpenCV is the most widely used software component library for computer vision. Initially used mainly for algorithm development and prototyping, in recent years OpenCV has also been used extensively for implementation and deployment of vision applications, including many mobile and embedded applications. Today, OpenCV runs on a wide range of operating systems including embedded Linux, Android, iOS, Windows Phone, and QNX.
Itseez, as OpenCV's primary maintainer, has been at the forefront of enabling OpenCV for embedded platforms and wants to share what it has learned. This talk will address several critical topics related to OpenCV in embedded systems, including cross-platform development best practices, performance profiling, benchmarking, and automated regression testing. Yury will present several real-world automotive use cases and the key lessons learned from them.
Distributed Caching in Kubernetes with HazelcastMesut Celik
As Monolith to Microservices migration almost became mainstream, Engineering Teams have to think about how their caching strategies will evolve in cloud-native world. Kubernetes is clear winner in containerized world so caching solutions must be cloud-ready and natural fit for Kubernetes.
Caching is an important piece in high performance microservices and choosing right architectural pattern can be crucial for your deployments. Hazelcast is a well known caching solution in open source community and can handle caching piece in microservices based applications.
In this talk, you will learn
* Distributed Caching With Hazelcast
* Distributed Caching Patterns in Kubernetes
* Kubernetes Deployment Options and Best Practices
* How to Handle Distributed Caching Day 2 Operations
Wix' internal ML Platform, whose mission is to allow data scientists and analysts at Wix to build, deploy, maintain, and monitor machine learning models in production with minimal engineering efforts
TensorRT is an NVIDIA tool that optimizes and accelerates deep learning models for production deployment. It performs optimizations like layer fusion, reduced precision from FP32 to FP16 and INT8, kernel auto-tuning, and multi-stream execution. These optimizations reduce latency and increase throughput. TensorRT automatically optimizes models by taking in a graph, performing optimizations, and outputting an optimized runtime engine.
RESTful Machine Learning with Flask and TensorFlow Serving - Carlo MazzaferroPyData
Those of us who use TensorFlow often focus on building the model that's most predictive, not the one that's most deployable. So how to put that hard work to work? In this talk, we'll walk through a strategy for taking your machine learning models from Jupyter Notebook into production and beyond.
What is Deep Learning | Deep Learning Simplified | Deep Learning Tutorial | E...Edureka!
This Edureka "What is Deep Learning" video will help you to understand about the relationship between Deep Learning, Machine Learning and Artificial Intelligence and how Deep Learning came into the picture. This tutorial will be discussing about Artificial Intelligence, Machine Learning and its limitations, how Deep Learning overcame Machine Learning limitations and different real-life applications of Deep Learning.
Below are the topics covered in this tutorial:
1. What Is Artificial Intelligence?
2. What Is Machine Learning?
3. Limitations Of Machine Learning
4. Deep Learning To The Rescue
5. What Is Deep Learning?
6. Deep Learning Applications
To take a structured training on Deep Learning, you can check complete details of our Deep Learning with TensorFlow course here: https://goo.gl/VeYiQZ
An introduction to the Transformers architecture and BERTSuman Debnath
The transformer is one of the most popular state-of-the-art deep (SOTA) learning architectures that is mostly used for natural language processing (NLP) tasks. Ever since the advent of the transformer, it has replaced RNN and LSTM for various tasks. The transformer also created a major breakthrough in the field of NLP and also paved the way for new revolutionary architectures such as BERT.
OpenGL is a powerful, low-level graphics toolkit with a steep learning curve that allows access to accelerated GPU hardware. Using OpenGL, developers achieve high fidelity, animated graphics ubiquitous in games, screen productions and scientific software.
Join us for a one-hour webinar and we will give a comprehensive overview of the many aspects of OpenGL development where Qt provides advanced interfaces that let the developer focus on the tasks instead of dealing with repetitive and error-prone, platform dependent issues.
This document provides an overview of ONNX and ONNX Runtime. ONNX is an open format for machine learning models that allows models to be shared across different frameworks and tools. ONNX Runtime is a cross-platform open source inference engine that runs ONNX models. It supports hardware acceleration and has a modular design that allows for custom operators and execution providers to extend its capabilities. The document discusses how ONNX helps with deploying machine learning models from research to production and how ONNX Runtime performs high performance inference through optimizations and hardware acceleration.
NVIDIA’s invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined modern computer graphics, and revolutionized parallel computing. More recently, GPU deep learning ignited modern AI — the next era of computing — with the GPU acting as the brain of computers, robots, and self-driving cars that can perceive and understand the world. Today, NVIDIA is increasingly known as “the AI computing company.”
Deep learning is a branch of machine learning that uses neural networks with multiple processing layers to learn representations of data with multiple levels of abstraction. It has been applied to problems like image recognition, natural language processing, and game playing. Deep learning architectures like deep neural networks use techniques like pretraining, dropout, and early stopping to avoid overfitting. Popular deep learning frameworks and libraries include TensorFlow, Keras, and PyTorch.
The RAPIDS suite of software libraries gives you the freedom to execute end-to-end data science and analytics pipelines entirely on GPUs. It relies on NVIDIA® CUDA® primitives for low-level compute optimization, but exposes that GPU parallelism and high-bandwidth memory speed through user-friendly Python interfaces.
Machine Learning. What is machine learning. Normal computer vs ML. Types of Machine Learning. Some ML Object detection methods. Faster CNN, RCNN, YOLO, SSD. Real Life ML Applications. Best Programming Languages for ML. Difference Between Machine Learning And Artificial Intelligence. Advantages of Machine Learning. Disadvantages of Machine Learning
While Go is the language-of-choice in the cloud-native world, Python has a huge community and makes it really easy to extend Kubernetes in only a few lines of code.
This talk shows examples on how to use Python to query the Kubernetes API, how to write simple controllers in only 10 lines of Python, how to build complete web UIs, and how to test everything with py.test and Kind.
Some of the open-source projects which will be covered: pykube-ng, Kubernetes Web View, kube-janitor, and Kopf (Kubernetes Operator Pythonic Framework).
Talk held in Prague on 2019-09-05:
https://www.meetup.com/Cloud-Native-Prague/events/263802447/
Slides reviewing the paper:
Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, and Illia Polosukhin. "Attention is all you need." In Advances in Neural Information Processing Systems, pp. 6000-6010. 2017.
The dominant sequence transduction models are based on complex recurrent orconvolutional neural networks in an encoder and decoder configuration. The best performing such models also connect the encoder and decoder through an attentionm echanisms. We propose a novel, simple network architecture based solely onan attention mechanism, dispensing with recurrence and convolutions entirely.Experiments on two machine translation tasks show these models to be superiorin quality while being more parallelizable and requiring significantly less timeto train. Our single model with 165 million parameters, achieves 27.5 BLEU onEnglish-to-German translation, improving over the existing best ensemble result by over 1 BLEU. On English-to-French translation, we outperform the previoussingle state-of-the-art with model by 0.7 BLEU, achieving a BLEU score of 41.1.
This document provides an introduction to Qt Quick, the declarative UI framework in Qt. It discusses key Qt Quick concepts like QML syntax, properties, binding, and user interface composition. It also covers cross-platform development with Qt, the Qt module system, and Qt UI technologies like Qt Quick, Qt Widgets and web/hybrid approaches. Finally, it discusses specific Qt Quick topics like elements, properties, identities, anchors, and user interaction handling.
Using MLOps to Bring ML to Production/The Promise of MLOpsWeaveworks
In this final Weave Online User Group of 2019, David Aronchick asks: have you ever struggled with having different environments to build, train and serve ML models, and how to orchestrate between them? While DevOps and GitOps have made huge traction in recent years, many customers struggle to apply these practices to ML workloads. This talk will focus on the ways MLOps has helped to effectively infuse AI into production-grade applications through establishing practices around model reproducibility, validation, versioning/tracking, and safe/compliant deployment. We will also talk about the direction for MLOps as an industry, and how we can use it to move faster, with more stability, than ever before.
The recording of this session is on our YouTube Channel here: https://youtu.be/twsxcwgB0ZQ
Speaker: David Aronchick, Head of Open Source ML Strategy, Microsoft
Bio: David leads Open Source Machine Learning Strategy at Azure. This means he spends most of his time helping humans to convince machines to be smarter. He is only moderately successful at this. Previously, David led product management for Kubernetes at Google, launched GKE, and co-founded the Kubeflow project. David has also worked at Microsoft, Amazon and Chef and co-founded three startups.
Sign up for a free Machine Learning Ops Workshop: http://bit.ly/MLOps_Workshop_List
Weaveworks will cover concepts such as GitOps (operations by pull request), Progressive Delivery (canary, A/B, blue-green), and how to apply those approaches to your machine learning operations to mitigate risk.
The objectives of the seminar are to shed a light on the premises of FP and give you a basic understanding of the pillars of FP so that you would feel enlightened at the end of the session. When you walk away from the seminar you should feel an inner light about the new way of programming and an urge & motivation to code like you never before did!
Functional programming should not be confused with imperative (or procedural) programming. Neither it is like object oriented programming. It is something different. Not radically so, since the concepts that we will be exploring are familiar programming concepts, just expressed in a different way. The philosophy behind how these concepts are applied to solving problems are also a little different. We shall learn and talk about essentially the fundamental elements of Functional Programming.
Slide 1
TypeScript
* This presentation is to show TypeScript's major feature and the benefit that it brings to your JavaScript projects.
* Our main objective is just to spark interest especially to those not familiar with the tool.
Slide 2
- What is TypeScript
* go to next slide
Slide 3
- Is a superset of JavaScript
* it simply means an extension to JavaScript
- Use JavaScript code on TypeScript
* JS code naturally works on TypeScript
* Which also means your beloved JavaScript libraries such as JQuery, or your fancy interacive plugins would work as well.
- TypeScript compiles to plain old JavaScript
* TS code compiles to simple and clean JS.
Slide 4
- Screenshot of TS compiled to JS
* In this example, compiling a TS class code would result to a JS version, and a regular JavaScript function when compiled is basically untouched.
Slide 5
- TypeScript's Main Feature
* So what does TS provide us with? What does it actually do?
Slide 6
- Static Type Checking
* TypeScript allows us to enable type checking by defining data types to your for ex. variables, function parameters and return types.
Slide 7
- Screenshot of basic Static Type Checking
* In this example…
* What I've done here was to assign supposedly wrong values for what the variables or parameters were meant to hold
* As JavaScript is a dynamic and untyped language these expressions would either fail or be okay when you run it on your browser.
* In TypeScript by enabling static type checking these potential errors are caught earlier (see the red marks on the expressions) and wouldn't even allow you to compile unless these are resolved.
* In addition you can also type arrays and object literals
Slide 8
- Effects of Static Type Checking
* As TS code is statically type-checked a side effect of such...
- Allows IDEs to perform live error checks
- Exposes auto-completion and code hinting
Slide 9
- Screenshot of code hinting
* Say I was coding JQuery on regular JavaScript code there would be no natural way to help me identify its class properties, methods and parameters... except through reading the API documentation or a separate plugin.
* As a result of static type checking this allows IDE's to access these class members as code hints
* So if this was a 3rd party library how much more if you are just referencing your own JavaScript/TypeScript files within your project.
Slide 10
- A few of the other cool features
* That was only the basic feature of TypeScript
* A few of the other cool features are...
Slide 11
- End
How to Get the Best Deep Learning performance with OpenVINO ToolkitYury Gorbachev
The document discusses Intel's OpenVINOTM toolkit, which provides tools to optimize deep learning models for deployment across Intel platforms. It notes that the toolkit has been adopted by over 30 customers and provides up to 5x faster and more efficient inference compared to frameworks like TensorFlow and MXNet. The toolkit allows models to run on CPU, GPU, VPU and FPGA with minimal code changes and supports post-training quantization and binarization. It includes tools for video preprocessing and benchmarks to measure application performance rather than just inference speed. The document recommends using the included high performance models from the Open Model Zoo and demo applications.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/itseez/embedded-vision-training/videos/pages/may-2015-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Yury Gorbachev, Principal Engineer at itseez, presents the "OpenCV for Embedded: Lessons Learned" tutorial at the May 2015 Embedded Vision Summit.
OpenCV is the most widely used software component library for computer vision. Initially used mainly for algorithm development and prototyping, in recent years OpenCV has also been used extensively for implementation and deployment of vision applications, including many mobile and embedded applications. Today, OpenCV runs on a wide range of operating systems including embedded Linux, Android, iOS, Windows Phone, and QNX.
Itseez, as OpenCV's primary maintainer, has been at the forefront of enabling OpenCV for embedded platforms and wants to share what it has learned. This talk will address several critical topics related to OpenCV in embedded systems, including cross-platform development best practices, performance profiling, benchmarking, and automated regression testing. Yury will present several real-world automotive use cases and the key lessons learned from them.
This document discusses lessons learned from using OpenCV for embedded vision applications. It notes that while OpenCV works well out of the box for desktop applications, embedded platforms present additional challenges like different processors, interfaces, and unpredictable performance. It recommends prototyping on desktop for faster development, then optimizing algorithms and porting to embedded hardware. Specific optimizations discussed include using ARMv8 processors, vendor-optimized OpenCV packages, and custom NEON-accelerated functions. An example product from Itseez that runs computer vision algorithms in real-time on ARM using these techniques is also presented.
Enabling Cross-platform Deep Learning Applications with Intel OpenVINO™Yury Gorbachev
The document discusses Intel's OpenVINOTM toolkit, which enables cross-platform deep learning applications. The toolkit provides a development environment for high performance computer vision and deep learning inference. It allows applications to run on multiple Intel accelerators with a single application binary and without retraining models. This allows for faster time to market, cross-platform portability, and future-proofing of applications.
Faster deep learning solutions from training to inference - Michele Tameni - ...Codemotion
Intel Deep Learning SDK enables using of optimized open source deep-learning frameworks, including Caffe and TensorFlow through a step-by-step wizard or iPython interactive notebooks. It includes easy and fast installation of all depended libraries and advanced tools for easy data pre-processing and model training, optimization and deployment, providing an end-to-end solution to the problem. In addition, it supports scale-out on multiple computers for training, as well as using compression methods for deployment of the models on various platforms, addressing memory and speed constraints.
Get Devops Training in Chennai with real-time experts at Besant Technologies, OMR. We believe that learning Devops with practical and theoretical will be the easiest way to understand the technology in quick manner. We designed this Devops from basic level to the latest advanced level
http://www.traininginsholinganallur.in/devops-training-in-chennai.html
This document summarizes Intel's Cluster Studio XE and Parallel Studio XE development tools. It provides an overview of the tools' capabilities for developing applications that can efficiently scale from few cores to many cores. The tools include compilers, performance profiling and analysis tools, and libraries that support parallel programming models and provide optimized math functions and algorithms.
Jython for embedded software validationPyCon Italia
This document discusses using Jython for embedded software validation. It proposes a runtime plugin model using Jython that allows dynamic code loading and multithreaded execution. This enables extending test automation capabilities and reusable test code. The model is implemented using Jython embedded in Eclipse, providing benefits like cross-platform code and dynamic reloading while avoiding risks of mixed language development.
Intel Parallel Studio XE 2016 網路開發工具包新版本功能介紹(現已上市,歡迎詢價) 購買 http://www.cheerchain.com.tw/ or http://www.appcenter.com.tw/
Distributed Tensorflow with Kubernetes - data2day - Jakob KaralusJakob Karalus
This document discusses using Distributed Tensorflow with Kubernetes for training neural networks. It covers:
- The need for distributed training to handle large datasets, deep models, and high accuracy requirements.
- Kubernetes as an orchestration tool for scheduling Tensorflow across nodes with GPUs.
- Key concepts like parameter servers, worker replicas, and synchronous/asynchronous training modes.
- Steps for setting up distributed Tensorflow jobs on Kubernetes including defining the cluster, assigning operations, creating training sessions, and packaging into containers.
- Considerations for enabling GPUs, building Docker images, writing deployments, and automating with tools like the Tensorflow Operator.
Multicore 101: Migrating Embedded Apps to Multicore with LinuxBrad Dixon
Multicore 101 provides an overview of migrating embedded applications to a multicore environment with Linux. It discusses the challenges in migrating large codebases of single-threaded legacy code to parallel multicore architectures. It then presents several proposed multicore solutions including combined hardware/software approaches using virtualization and hypervisors. The presentation recommends containing, exploiting, analyzing and optimizing applications as the pathway for migrating code and emphasizes the importance of using the right tools to facilitate the migration. It promotes evaluating MontaVista TestDrive on Freescale multicore platforms as a way to assess multicore solutions.
Develop, Deploy, and Innovate with Intel® Cluster ReadyIntel IT Center
The document discusses Intel's Cluster Ready specification and Intel Cluster Checker tool. The specification has been updated to version 1.3 to add support for Intel Xeon Phi coprocessors. The Cluster Checker tool has also been updated to version 2.1 to verify compliance with the new specification and support Intel Xeon Phi based clusters. The updates allow for easier development, deployment, and management of high performance Intel-based computer clusters.
This document discusses Juniper's automation tools and capabilities. It provides an overview of Juniper's automation offerings, including tools for build (provisioning), configure (configuration), and collect (monitoring) phases. Example use cases for enterprise IT and cloud automation are also presented. Competitive advantages over Cisco and Arista are highlighted, such as Juniper providing a common set of automation tools across all its products and rich off-box functionality using Python libraries.
OS for AI: Elastic Microservices & the Next Gen of MLNordic APIs
AI has been a hot topic lately, with advances being made constantly in what is possible, there has not been as much discussion of the infrastructure and scaling challenges that come with it. How do you support dozens of different languages and frameworks, and make them interoperate invisibly? How do you scale to run abstract code from thousands of different developers, simultaneously and elastically, while maintaining less than 15ms of overhead?
At Algorithmia, we’ve built, deployed, and scaled thousands of algorithms and machine learning models, using every kind of framework (from scikit-learn to tensorflow). We’ve seen many of the challenges faced in this area, and in this talk I’ll share some insights into the problems you’re likely to face, and how to approach solving them.
In brief, we’ll examine the need for, and implementations of, a complete “Operating System for AI” – a common interface for different algorithms to be used and combined, and a general architecture for serverless machine learning which is discoverable, versioned, scalable and sharable.
Hyper-threading technology allows a single processor to appear and function as multiple processors. It was first implemented in 2002 in Intel Pentium 4 Xeon processors. Hyper-threading works by dividing processor workload into threads that can be executed concurrently on different processor execution units. This makes more efficient use of processor resources and improves performance on multi-threaded software. While it increases throughput, shared processor resources between threads can also lead to conflicts that result in reduced performance in some cases.
The journey to Native Cloud Architecture & Microservices, tracing the footste...Mek Srunyu Stittri
The document discusses Netflix's adoption of microservices and continuous delivery to improve speed and agility. Key points include:
1) Netflix moved to microservices and continuous delivery on the cloud to dramatically speed up product development and deployment.
2) This allowed independent teams to deploy code frequently without coordination, with automated testing and deployment replacing handoffs and long release cycles.
3) Netflix's approach involved building stateless, independently deployable microservices; continuous monitoring; and other techniques to enable developers to deploy code safely and rapidly.
Get the Facts: Oracle's Unbreakable Enterprise KernelTerry Wang
1) Oracle introduced the Unbreakable Enterprise Kernel for Oracle Linux, which is optimized for Oracle software and provides significant performance gains over the Red Hat compatible kernel.
2) The Unbreakable Enterprise Kernel includes many new features like improved power management, data integrity, and diagnostic tools.
3) Oracle recommends customers use the Unbreakable Enterprise Kernel for all Oracle software on Linux, though it will continue to support the Red Hat compatible kernel.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/altera/embedded-vision-training/videos/pages/may-2015-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Deshanand Singh, Director of Software Engineering at Altera, presents the "Efficient Implementation of Convolutional Neural Networks using OpenCL on FPGAs" tutorial at the May 2015 Embedded Vision Summit.
Convolutional neural networks (CNN) are becoming increasingly popular in embedded applications such as vision processing and automotive driver assistance systems. The structure of CNN systems is characterized by cascades of FIR filters and transcendental functions. FPGA technology offers a very efficient way of implementing these structures by allowing designers to build custom hardware datapaths that implement the CNN structure. One challenge of using FPGAs revolves around the design flow that has been traditionally centered around tedious hardware description languages.
In this talk, Deshanand gives a detailed explanation of how CNN algorithms can be expressed in OpenCL and compiled directly to FPGA hardware. He gives detail on code optimizations and provides comparisons with the efficiency of hand-coded implementations.
“Microservices” have become a trendy development strategy. Hosting and running such services used to be pretty painful... but here comes Service Fabric! Let’s take a closer look at this platform, its different development models and all the features it offers, and not only for microservices!
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdfAbi john
Analyze the growth of meme coins from mere online jokes to potential assets in the digital economy. Explore the community, culture, and utility as they elevate themselves to a new era in cryptocurrency.
Complete Guide to Advanced Logistics Management Software in Riyadh.pdfSoftware Company
Explore the benefits and features of advanced logistics management software for businesses in Riyadh. This guide delves into the latest technologies, from real-time tracking and route optimization to warehouse management and inventory control, helping businesses streamline their logistics operations and reduce costs. Learn how implementing the right software solution can enhance efficiency, improve customer satisfaction, and provide a competitive edge in the growing logistics sector of Riyadh.
AI Changes Everything – Talk at Cardiff Metropolitan University, 29th April 2...Alan Dix
Talk at the final event of Data Fusion Dynamics: A Collaborative UK-Saudi Initiative in Cybersecurity and Artificial Intelligence funded by the British Council UK-Saudi Challenge Fund 2024, Cardiff Metropolitan University, 29th April 2025
https://alandix.com/academic/talks/CMet2025-AI-Changes-Everything/
Is AI just another technology, or does it fundamentally change the way we live and think?
Every technology has a direct impact with micro-ethical consequences, some good, some bad. However more profound are the ways in which some technologies reshape the very fabric of society with macro-ethical impacts. The invention of the stirrup revolutionised mounted combat, but as a side effect gave rise to the feudal system, which still shapes politics today. The internal combustion engine offers personal freedom and creates pollution, but has also transformed the nature of urban planning and international trade. When we look at AI the micro-ethical issues, such as bias, are most obvious, but the macro-ethical challenges may be greater.
At a micro-ethical level AI has the potential to deepen social, ethnic and gender bias, issues I have warned about since the early 1990s! It is also being used increasingly on the battlefield. However, it also offers amazing opportunities in health and educations, as the recent Nobel prizes for the developers of AlphaFold illustrate. More radically, the need to encode ethics acts as a mirror to surface essential ethical problems and conflicts.
At the macro-ethical level, by the early 2000s digital technology had already begun to undermine sovereignty (e.g. gambling), market economics (through network effects and emergent monopolies), and the very meaning of money. Modern AI is the child of big data, big computation and ultimately big business, intensifying the inherent tendency of digital technology to concentrate power. AI is already unravelling the fundamentals of the social, political and economic world around us, but this is a world that needs radical reimagining to overcome the global environmental and human challenges that confront us. Our challenge is whether to let the threads fall as they may, or to use them to weave a better future.
AI EngineHost Review: Revolutionary USA Datacenter-Based Hosting with NVIDIA ...SOFTTECHHUB
I started my online journey with several hosting services before stumbling upon Ai EngineHost. At first, the idea of paying one fee and getting lifetime access seemed too good to pass up. The platform is built on reliable US-based servers, ensuring your projects run at high speeds and remain safe. Let me take you step by step through its benefits and features as I explain why this hosting solution is a perfect fit for digital entrepreneurs.
TrsLabs - Fintech Product & Business ConsultingTrs Labs
Hybrid Growth Mandate Model with TrsLabs
Strategic Investments, Inorganic Growth, Business Model Pivoting are critical activities that business don't do/change everyday. In cases like this, it may benefit your business to choose a temporary external consultant.
An unbiased plan driven by clearcut deliverables, market dynamics and without the influence of your internal office equations empower business leaders to make right choices.
Getting things done within a budget within a timeframe is key to Growing Business - No matter whether you are a start-up or a big company
Talk to us & Unlock the competitive advantage
Increasing Retail Store Efficiency How can Planograms Save Time and Money.pptxAnoop Ashok
In today's fast-paced retail environment, efficiency is key. Every minute counts, and every penny matters. One tool that can significantly boost your store's efficiency is a well-executed planogram. These visual merchandising blueprints not only enhance store layouts but also save time and money in the process.
Massive Power Outage Hits Spain, Portugal, and France: Causes, Impact, and On...Aqusag Technologies
In late April 2025, a significant portion of Europe, particularly Spain, Portugal, and parts of southern France, experienced widespread, rolling power outages that continue to affect millions of residents, businesses, and infrastructure systems.
Noah Loul Shares 5 Steps to Implement AI Agents for Maximum Business Efficien...Noah Loul
Artificial intelligence is changing how businesses operate. Companies are using AI agents to automate tasks, reduce time spent on repetitive work, and focus more on high-value activities. Noah Loul, an AI strategist and entrepreneur, has helped dozens of companies streamline their operations using smart automation. He believes AI agents aren't just tools—they're workers that take on repeatable tasks so your human team can focus on what matters. If you want to reduce time waste and increase output, AI agents are the next move.
This is the keynote of the Into the Box conference, highlighting the release of the BoxLang JVM language, its key enhancements, and its vision for the future.
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungenpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-nomad-web-best-practices-und-verwaltung-von-multiuser-umgebungen/
HCL Nomad Web wird als die nächste Generation des HCL Notes-Clients gefeiert und bietet zahlreiche Vorteile, wie die Beseitigung des Bedarfs an Paketierung, Verteilung und Installation. Nomad Web-Client-Updates werden “automatisch” im Hintergrund installiert, was den administrativen Aufwand im Vergleich zu traditionellen HCL Notes-Clients erheblich reduziert. Allerdings stellt die Fehlerbehebung in Nomad Web im Vergleich zum Notes-Client einzigartige Herausforderungen dar.
Begleiten Sie Christoph und Marc, während sie demonstrieren, wie der Fehlerbehebungsprozess in HCL Nomad Web vereinfacht werden kann, um eine reibungslose und effiziente Benutzererfahrung zu gewährleisten.
In diesem Webinar werden wir effektive Strategien zur Diagnose und Lösung häufiger Probleme in HCL Nomad Web untersuchen, einschließlich
- Zugriff auf die Konsole
- Auffinden und Interpretieren von Protokolldateien
- Zugriff auf den Datenordner im Cache des Browsers (unter Verwendung von OPFS)
- Verständnis der Unterschiede zwischen Einzel- und Mehrbenutzerszenarien
- Nutzung der Client Clocking-Funktion
Procurement Insights Cost To Value Guide.pptxJon Hansen
Procurement Insights integrated Historic Procurement Industry Archives, serves as a powerful complement — not a competitor — to other procurement industry firms. It fills critical gaps in depth, agility, and contextual insight that most traditional analyst and association models overlook.
Learn more about this value- driven proprietary service offering here.
Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...Impelsys Inc.
Impelsys provided a robust testing solution, leveraging a risk-based and requirement-mapped approach to validate ICU Connect and CritiXpert. A well-defined test suite was developed to assess data communication, clinical data collection, transformation, and visualization across integrated devices.
What is Model Context Protocol(MCP) - The new technology for communication bw...Vishnu Singh Chundawat
The MCP (Model Context Protocol) is a framework designed to manage context and interaction within complex systems. This SlideShare presentation will provide a detailed overview of the MCP Model, its applications, and how it plays a crucial role in improving communication and decision-making in distributed systems. We will explore the key concepts behind the protocol, including the importance of context, data management, and how this model enhances system adaptability and responsiveness. Ideal for software developers, system architects, and IT professionals, this presentation will offer valuable insights into how the MCP Model can streamline workflows, improve efficiency, and create more intuitive systems for a wide range of use cases.
Dev Dives: Automate and orchestrate your processes with UiPath MaestroUiPathCommunity
This session is designed to equip developers with the skills needed to build mission-critical, end-to-end processes that seamlessly orchestrate agents, people, and robots.
📕 Here's what you can expect:
- Modeling: Build end-to-end processes using BPMN.
- Implementing: Integrate agentic tasks, RPA, APIs, and advanced decisioning into processes.
- Operating: Control process instances with rewind, replay, pause, and stop functions.
- Monitoring: Use dashboards and embedded analytics for real-time insights into process instances.
This webinar is a must-attend for developers looking to enhance their agentic automation skills and orchestrate robust, mission-critical processes.
👨🏫 Speaker:
Andrei Vintila, Principal Product Manager @UiPath
This session streamed live on April 29, 2025, 16:00 CET.
Check out all our upcoming Dev Dives sessions at https://community.uipath.com/dev-dives-automation-developer-2025/.
Linux Support for SMARC: How Toradex Empowers Embedded DevelopersToradex
Toradex brings robust Linux support to SMARC (Smart Mobility Architecture), ensuring high performance and long-term reliability for embedded applications. Here’s how:
• Optimized Torizon OS & Yocto Support – Toradex provides Torizon OS, a Debian-based easy-to-use platform, and Yocto BSPs for customized Linux images on SMARC modules.
• Seamless Integration with i.MX 8M Plus and i.MX 95 – Toradex SMARC solutions leverage NXP’s i.MX 8 M Plus and i.MX 95 SoCs, delivering power efficiency and AI-ready performance.
• Secure and Reliable – With Secure Boot, over-the-air (OTA) updates, and LTS kernel support, Toradex ensures industrial-grade security and longevity.
• Containerized Workflows for AI & IoT – Support for Docker, ROS, and real-time Linux enables scalable AI, ML, and IoT applications.
• Strong Ecosystem & Developer Support – Toradex offers comprehensive documentation, developer tools, and dedicated support, accelerating time-to-market.
With Toradex’s Linux support for SMARC, developers get a scalable, secure, and high-performance solution for industrial, medical, and AI-driven applications.
Do you have a specific project or application in mind where you're considering SMARC? We can help with Free Compatibility Check and help you with quick time-to-market
For more information: https://www.toradex.com/computer-on-modules/smarc-arm-family
2. Intel ConfidentialIntel Confidential
Brief OpenVINO™ Introduction
• OpenVINO ™ is
• set of tools and libraries for CV/DL application developers
• high performance, low footprint solution for deployment
• API for unified access to CV/DL capabilities of Intel platforms
• OpenVINO ™ is not
• tool for data scientists
• solution for training of deep learning models
3. Intel ConfidentialIntel Confidential 3
OpenVINO™ Benefits
• Powerful combination of highly optimized Classical CV and DL primitives
• Allows to run inference on Intel CPU, GPU, VPU and FPGA
• Best performing solution on all Intel architectures
• ~2x faster than fastest TensorFlow and MxNet on CPU
• Smallest execution footprint (lowest memory consumption)
• 2x smaller than MXNet, 4x smaller than TensorFlow
• Minimum # of dependencies (no dependencies on training frameworks)
• Multiple OS support (Linux and Windows)
4. Intel ConfidentialIntel Confidential 4
OpenVINO™ Specifics
• Highly optimized implementations of DL primitives
• Most efficient on each Intel platform
• Focus on inference only
• Aggressive layer fusion at the inference step
• Including HW accelerated steps tuned for inference
• Efficient activation memory reuse
• Often close to bare minimum
5. Intel ConfidentialIntel Confidential 5
DL Workflow
Caffe
MXNet
TensorFlow
Caffe2
PyTorch
Serialized
trained
DL model
ONNX
MKLDNN
Plugin
clDNN
Plugin
FPGA
Plugin
Myriad
Plugin
Inference
Engine
Deploy
Application
Model
Optimizer
IR
.xml
.bin
Step 1: Import model from Framework format
to Framework independent representation
Step 2: Update application
to use Inference Engine API
Eliminate unnecessary layers,
lossless fusion where possible
Remove framework dependency
Accuracy against original model ensured
6. Intel ConfidentialIntel Confidential 6
Customization Capabilities
• OpenVINO™ provides good coverage of DL primitives out of the box
• Constantly growing list of primitives to support new DL topologies
• Frequent releases, substantial additions
• Not a problem if something is missing!
• Good extension mechanism for adding new primitives
• Possible to add proprietary layers, more optimized layers, etc.
• Both in Model Optimizer and Inference Engine (import and run)
7. Intel ConfidentialIntel Confidential 7
Application Design Workflow
Desig
n
Verify
logic
Debug
Fix
Best done on CPU:
- Easier to verify
- Simpler debugging procedures
CHANGE
TARGET
Check
scalability
Fix
pipeline
System
testing
Best done on Actual target (e.g. VPU):
- Exact performance
- Correct timings
Accuracy and functionality
across targets
8. Intel ConfidentialIntel Confidential 8
Heterogeneous Execution
• When a certain primitive is not supported on a target
• Custom proprietary primitive or inefficient HW for a task
• Heterogeneous execution ensures full topology execution
• Automatic data transfer between targets whenever needed
• Work splitting and scheduling
• No need to do any manual network manipulations!
FPGA CPU
9. Intel ConfidentialIntel Confidential 9
Power of Parallel Execution
Asynchronous API provides capabilities for:
• Running main thread in parallel with ongoing inference (CPU/GPU)
• Hiding data transfer latency for accelerators (VPU, FPGA)
• Filling accelerators with work
Transfer 1
Inference 1
Transfer 2
Transfer 1 Transfer 2
Inference 2
Inference 1 Inference 2
Result 1
Result 1
Result 2
Result 2
Sync API
Async API
10. Intel ConfidentialIntel Confidential 10
Efficient Frame Preprocessing
DECODE
Layout
Transform
Resize to
network
Detection
network
Resize to
network
Object Analysis
Network
Crop
Preprocessing
DL Inference
• Preprocessing is typically done via manual coding or using libraries
• OpenCV is most popular
• Fastest OpenCV build is available in OpenVINO™
• Deep Learning Inference Engine encapsulates basic preprocessing capability
• Automatic frame resize based on input frame and network size
• Automatic layout conversion and cropping
• In a nutshell –> Just provide frame and it will be suitable for inference automatically
11. Intel ConfidentialIntel Confidential 11
INT8 Quantization for CPU
• INT8 provides additional acceleration using AVX-512
• Not all CPU targets support it
• Very minor quality/accuracy loss
• No retraining is required
• No code update is needed
IR
Dataset
Analysis
tool Updated IR
12. Intel ConfidentialIntel Confidential 12
OpenVINO™ Model Zoo
• OpenVINO provides pre-trained DL models for deployment
• Lightweight, low compute, real time on Intel platforms
• Cover popular CV use cases
• Face analysis, Security use-cases (person, vehicle, bicycle detection)
• Transportation analytics (road segmentation, vehicle/pedestrian detection)
13. Intel ConfidentialIntel Confidential 13
Extensive Set of Samples
• Actual examples of applications, not just API demonstration
• Switching between targets, work distribution
• Multiple models in pipeline with preprocessing
• Open Models Zoo demonstration