For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2018-embedded-vision-summit-sze
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Vivienne Sze, Associate Professor at MIT, presents the "Approaches for Energy Efficient Implementation of Deep Neural Networks" tutorial at the May 2018 Embedded Vision Summit.
Deep neural networks (DNNs) are proving very effective for a variety of challenging machine perception tasks. But these algorithms are very computationally demanding. To enable DNNs to be used in practical applications, it’s critical to find efficient ways to implement them.
This talk explores how DNNs are being mapped onto today’s processor architectures, and how these algorithms are evolving to enable improved efficiency. Sze explores the energy consumption of commonly used CNNs versus their accuracy, and provides insights on "energy-aware" pruning of these networks.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2021/09/introduction-to-dnn-model-compression-techniques-a-presentation-from-xailient/
Sabina Pokhrel, Customer Success AI Engineer at Xailient, presents the “Introduction to DNN Model Compression Techniques” tutorial at the May 2021 Embedded Vision Summit.
Embedding real-time large-scale deep learning vision applications at the edge is challenging due to their huge computational, memory, and bandwidth requirements. System architects can mitigate these demands by modifying deep-neural networks to make them more energy efficient and less demanding of processing resources by applying various model compression approaches.
In this talk, Pokhrel provides an introduction to four established techniques for model compression. She discusses network pruning, quantization, knowledge distillation and low-rank factorization compression approaches.
introduction to deep Learning with full detailsonykhan3
1. Deep learning involves using neural networks with multiple hidden layers to learn representations of data with multiple levels of abstraction.
2. These neural networks are able to learn increasingly complex features from the input data as the number of layers increases. The layers closer to the input learn simpler features while layers further from the input learn complex patterns in the data.
3. A breakthrough in deep learning was developing algorithms that can successfully train deep neural networks by unsupervised learning on each layer before using the learned features for supervised learning on the final layer. This pretraining helps the network learn useful internal representations.
It’s long ago, approx. 30 years, since AI was not only a topic for Science-Fiction writers, but also a major research field surrounded with huge hopes and investments. But the over-inflated expectations ended in a subsequent crash and followed by a period of absent funding and interest – the so-called AI winter. However, the last 3 years changed everything – again. Deep learning, a machine learning technique inspired by the human brain, successfully crushed one benchmark after another and tech companies, like Google, Facebook and Microsoft, started to invest billions in AI research. “The pace of progress in artificial general intelligence is incredible fast” (Elon Musk – CEO Tesla & SpaceX) leading to an AI that “would be either the best or the worst thing ever to happen to humanity” (Stephen Hawking – Physicist).
What sparked this new Hype? How is Deep Learning different from previous approaches? Are the advancing AI technologies really a threat for humanity? Let’s look behind the curtain and unravel the reality. This talk will explore why Sundar Pichai (CEO Google) recently announced that “machine learning is a core transformative way by which Google is rethinking everything they are doing” and explain why "Deep Learning is probably one of the most exciting things that is happening in the computer industry” (Jen-Hsun Huang – CEO NVIDIA).
Either a new AI “winter is coming” (Ned Stark – House Stark) or this new wave of innovation might turn out as the “last invention humans ever need to make” (Nick Bostrom – AI Philosoph). Or maybe it’s just another great technology helping humans to achieve more.
The document discusses Long Short Term Memory (LSTM) networks, which are a type of recurrent neural network capable of learning long-term dependencies. It explains that unlike standard RNNs, LSTMs use forget, input, and output gates to control the flow of information into and out of the cell state, allowing them to better capture long-range temporal dependencies in sequential data like text, audio, and time-series data. The document provides details on how LSTM gates work and how LSTMs can be used for applications involving sequential data like machine translation and question answering.
1) Deep learning is a type of machine learning that uses neural networks with many layers to learn representations of data with multiple levels of abstraction.
2) Deep learning techniques include unsupervised pretrained networks, convolutional neural networks, recurrent neural networks, and recursive neural networks.
3) The advantages of deep learning include automatic feature extraction from raw data with minimal human effort, and surpassing conventional machine learning algorithms in accuracy across many data types.
Model Compression (NanheeKim)
@NanheeKim @nh9k
질문이 있으면 언제든지 연락주세요!
공부한 것을 바탕으로 작성한 ppt입니다.
출처는 슬라이드 마지막에 있습니다!
Please, feel free to contact me, if you have any questions!
github: https://github.com/nh9k
email: [email protected]
Transfer Learning and Fine-tuning Deep Neural NetworksPyData
This document outlines Anusua Trivedi's talk on transfer learning and fine-tuning deep neural networks. The talk covers traditional machine learning versus deep learning, using deep convolutional neural networks (DCNNs) for image analysis, transfer learning and fine-tuning DCNNs, recurrent neural networks (RNNs), and case studies applying these techniques to diabetic retinopathy prediction and fashion image caption generation.
- The document discusses trends in AI chips, including the rise of deep learning models enabled by increased computing power and data availability.
- It outlines the AI stack from algorithms and neural network models down to chips, memory, and hardware. Popular deep learning model types and applications are also summarized.
- The trends are towards more specialized hardware like Google's TPUs for cloud servers and dedicated chips for mobile/edge devices from companies like Qualcomm and Nvidia. Processing-in-memory and new memory technologies may help address bandwidth bottlenecks.
- Overall hardware is still catching up to the needs of large neural networks, and there is a lack of unified software tools and frameworks to program diverse AI accelerators.
Spiking neural network: an introduction IDalin Zhang
1) The document discusses Spiking Neural Networks (SNNs), which are a type of neural network that more closely mimic biological neural behavior.
2) It describes the Leaky Integrate-and-Fire (LIF) neuron model, which is commonly used in SNNs. The LIF model integrates inputs over time and generates spikes when the voltage exceeds a threshold.
3) Different encoding approaches are discussed for representing input data as spike trains, including rate coding, temporal coding, population coding, and the Hough Spiker Algorithm. These approaches transform real-valued inputs into spike timings.
Introduction to Recurrent Neural NetworkKnoldus Inc.
The document provides an introduction to recurrent neural networks (RNNs). It discusses how RNNs differ from feedforward neural networks in that they have internal memory and can use their output from the previous time step as input. This allows RNNs to process sequential data like time series. The document outlines some common RNN types and explains the vanishing gradient problem that can occur in RNNs due to multiplication of small gradient values over many time steps. It discusses solutions to this problem like LSTMs and techniques like weight initialization and gradient clipping.
Faster R-CNN improves object detection by introducing a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. The RPN slides over feature maps and predicts object bounds and objectness at each position. During training, anchors are assigned positive or negative labels based on Intersection over Union with ground truth boxes. Faster R-CNN runs the RPN in parallel with Fast R-CNN for detection, end-to-end in a single network and stage. This achieves state-of-the-art object detection speed and accuracy while eliminating computationally expensive selective search for proposals.
Image classification with Deep Neural NetworksYogendra Tamang
This document discusses image classification using deep neural networks. It provides background on image classification and convolutional neural networks. The document outlines techniques like activation functions, pooling, dropout and data augmentation to prevent overfitting. It summarizes a paper on ImageNet classification using CNNs with multiple convolutional and fully connected layers. The paper achieved state-of-the-art results on ImageNet in 2010 and 2012 by training CNNs on a large dataset using multiple GPUs.
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)UMBC
We trained a large, deep convolutional neural network to classify the 1.2 million
high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 dif-
ferent classes. On the test data, we achieved top-1 and top-5 error rates of 37.5%
and 17.0% which is considerably better than the previous state-of-the-art. The
neural network, which has 60 million parameters and 650,000 neurons, consists
of five convolutional layers, some of which are followed by max-pooling layers,
and three fully-connected layers with a final 1000-way softmax. To make train-
ing faster, we used non-saturating neurons and a very efficient GPU implemen-
tation of the convolution operation. To reduce overfitting in the fully-connected
layers we employed a recently-developed regularization method called “dropout”
that proved to be very effective. We also entered a variant of this model in the
ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%,
compared to 26.2% achieved by the second-best entry.
Hardware for deep learning includes CPUs, GPUs, FPGAs, and ASICs. CPUs are general purpose but support deep learning through instructions like AVX-512 and libraries. GPUs like NVIDIA and AMD models are commonly used due to high parallelism and memory bandwidth. FPGAs offer high efficiency but require specialized programming. ASICs like Google's TPU are customized for deep learning and provide high performance but limited flexibility. Emerging hardware aims to improve efficiency and better match neural network computations.
The document discusses computer vision with deep learning. It provides an overview of convolutional neural networks and their use in computer vision applications like image classification and object detection. Specifically, it discusses how CNNs use convolutional layers to learn visual features from images and provide examples of CNNs being used for pipeline defect classification and filler cap quality control.
Deep learning is a type of machine learning that uses neural networks with multiple layers between the input and output layers. It allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. Deep learning has achieved great success in computer vision, speech recognition, and natural language processing due to recent advances in algorithms, computing power, and the availability of large datasets. Deep learning models can learn complex patterns directly from large amounts of unlabeled data without relying on human-engineered features.
FPGA Hardware Accelerator for Machine Learning
Machine learning publications and models are growing exponentially, outpacing Moore's law. Hardware acceleration using FPGAs, GPUs, and ASICs can provide performance gains over CPU-only implementations for machine learning workloads. FPGAs allow for reprogramming after manufacturing and can accelerate parts of machine learning algorithms through customized hardware while sharing computations between the FPGA and CPU. Vitis AI is a software stack that optimizes machine learning models for deployment on Xilinx FPGAs, providing pre-optimized models, tools for optimization and quantization, and high-level APIs.
This document provides an overview of deep learning, including its history, algorithms, tools, and applications. It begins with the history and evolution of deep learning techniques. It then discusses popular deep learning algorithms like convolutional neural networks, recurrent neural networks, autoencoders, and deep reinforcement learning. It also covers commonly used tools for deep learning and highlights applications in areas such as computer vision, natural language processing, and games. In the end, it discusses the future outlook and opportunities of deep learning.
Recurrent Neural Networks are popular Deep Learning models that have shown great promise to achieve state-of-the-art results in many tasks like Computer Vision, NLP, Finance and much more. Although being models proposed several years ago, RNN have gained popularity recently. In this talk, we will review how these models evolved over the years, dissection of RNN, current applications and its future.
This document provides an agenda for a presentation on deep learning, neural networks, convolutional neural networks, and interesting applications. The presentation will include introductions to deep learning and how it differs from traditional machine learning by learning feature representations from data. It will cover the history of neural networks and breakthroughs that enabled training of deeper models. Convolutional neural network architectures will be overviewed, including convolutional, pooling, and dense layers. Applications like recommendation systems, natural language processing, and computer vision will also be discussed. There will be a question and answer section.
This document summarizes Melanie Swan's presentation on deep learning. It began with defining key deep learning concepts and techniques, including neural networks, supervised vs. unsupervised learning, and convolutional neural networks. It then explained how deep learning works by using multiple processing layers to extract higher-level features from data and make predictions. Deep learning has various applications like image recognition and speech recognition. The presentation concluded by discussing how deep learning is inspired by concepts from physics and statistical mechanics.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session.
Artificial Neural Network seminar presentation using ppt.Mohd Faiz
- Artificial neural networks are inspired by biological neural networks and learning processes. They attempt to mimic the workings of the brain using simple units called artificial neurons that are connected in networks.
- Learning in neural networks involves modifying the synaptic strengths between neurons through mathematical optimization techniques. The goal is to minimize an error function that measures how well the network can approximate or complete a task.
- Neural networks can learn complex nonlinear functions through training algorithms like backpropagation that determine how to adjust the synaptic weights to improve performance on the learning task.
DLD meetup 2017, Efficient Deep LearningBrodmann17
The document discusses efficient techniques for deep learning on edge devices. It begins by noting that deep neural networks have high computational complexity which makes inference inefficient for edge devices without powerful GPUs. It then outlines the deep learning stack from hardware to libraries to frameworks to algorithms. The document focuses on how algorithms define model complexity and discusses the evolution of CNN architectures from LeNet5 to ResNet which generally increased in complexity. It covers techniques for reducing model size and operations like pruning, quantization, and knowledge distillation. The challenges of real-life applications on edge devices are discussed.
- The document discusses trends in AI chips, including the rise of deep learning models enabled by increased computing power and data availability.
- It outlines the AI stack from algorithms and neural network models down to chips, memory, and hardware. Popular deep learning model types and applications are also summarized.
- The trends are towards more specialized hardware like Google's TPUs for cloud servers and dedicated chips for mobile/edge devices from companies like Qualcomm and Nvidia. Processing-in-memory and new memory technologies may help address bandwidth bottlenecks.
- Overall hardware is still catching up to the needs of large neural networks, and there is a lack of unified software tools and frameworks to program diverse AI accelerators.
Spiking neural network: an introduction IDalin Zhang
1) The document discusses Spiking Neural Networks (SNNs), which are a type of neural network that more closely mimic biological neural behavior.
2) It describes the Leaky Integrate-and-Fire (LIF) neuron model, which is commonly used in SNNs. The LIF model integrates inputs over time and generates spikes when the voltage exceeds a threshold.
3) Different encoding approaches are discussed for representing input data as spike trains, including rate coding, temporal coding, population coding, and the Hough Spiker Algorithm. These approaches transform real-valued inputs into spike timings.
Introduction to Recurrent Neural NetworkKnoldus Inc.
The document provides an introduction to recurrent neural networks (RNNs). It discusses how RNNs differ from feedforward neural networks in that they have internal memory and can use their output from the previous time step as input. This allows RNNs to process sequential data like time series. The document outlines some common RNN types and explains the vanishing gradient problem that can occur in RNNs due to multiplication of small gradient values over many time steps. It discusses solutions to this problem like LSTMs and techniques like weight initialization and gradient clipping.
Faster R-CNN improves object detection by introducing a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. The RPN slides over feature maps and predicts object bounds and objectness at each position. During training, anchors are assigned positive or negative labels based on Intersection over Union with ground truth boxes. Faster R-CNN runs the RPN in parallel with Fast R-CNN for detection, end-to-end in a single network and stage. This achieves state-of-the-art object detection speed and accuracy while eliminating computationally expensive selective search for proposals.
Image classification with Deep Neural NetworksYogendra Tamang
This document discusses image classification using deep neural networks. It provides background on image classification and convolutional neural networks. The document outlines techniques like activation functions, pooling, dropout and data augmentation to prevent overfitting. It summarizes a paper on ImageNet classification using CNNs with multiple convolutional and fully connected layers. The paper achieved state-of-the-art results on ImageNet in 2010 and 2012 by training CNNs on a large dataset using multiple GPUs.
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)UMBC
We trained a large, deep convolutional neural network to classify the 1.2 million
high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 dif-
ferent classes. On the test data, we achieved top-1 and top-5 error rates of 37.5%
and 17.0% which is considerably better than the previous state-of-the-art. The
neural network, which has 60 million parameters and 650,000 neurons, consists
of five convolutional layers, some of which are followed by max-pooling layers,
and three fully-connected layers with a final 1000-way softmax. To make train-
ing faster, we used non-saturating neurons and a very efficient GPU implemen-
tation of the convolution operation. To reduce overfitting in the fully-connected
layers we employed a recently-developed regularization method called “dropout”
that proved to be very effective. We also entered a variant of this model in the
ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%,
compared to 26.2% achieved by the second-best entry.
Hardware for deep learning includes CPUs, GPUs, FPGAs, and ASICs. CPUs are general purpose but support deep learning through instructions like AVX-512 and libraries. GPUs like NVIDIA and AMD models are commonly used due to high parallelism and memory bandwidth. FPGAs offer high efficiency but require specialized programming. ASICs like Google's TPU are customized for deep learning and provide high performance but limited flexibility. Emerging hardware aims to improve efficiency and better match neural network computations.
The document discusses computer vision with deep learning. It provides an overview of convolutional neural networks and their use in computer vision applications like image classification and object detection. Specifically, it discusses how CNNs use convolutional layers to learn visual features from images and provide examples of CNNs being used for pipeline defect classification and filler cap quality control.
Deep learning is a type of machine learning that uses neural networks with multiple layers between the input and output layers. It allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. Deep learning has achieved great success in computer vision, speech recognition, and natural language processing due to recent advances in algorithms, computing power, and the availability of large datasets. Deep learning models can learn complex patterns directly from large amounts of unlabeled data without relying on human-engineered features.
FPGA Hardware Accelerator for Machine Learning
Machine learning publications and models are growing exponentially, outpacing Moore's law. Hardware acceleration using FPGAs, GPUs, and ASICs can provide performance gains over CPU-only implementations for machine learning workloads. FPGAs allow for reprogramming after manufacturing and can accelerate parts of machine learning algorithms through customized hardware while sharing computations between the FPGA and CPU. Vitis AI is a software stack that optimizes machine learning models for deployment on Xilinx FPGAs, providing pre-optimized models, tools for optimization and quantization, and high-level APIs.
This document provides an overview of deep learning, including its history, algorithms, tools, and applications. It begins with the history and evolution of deep learning techniques. It then discusses popular deep learning algorithms like convolutional neural networks, recurrent neural networks, autoencoders, and deep reinforcement learning. It also covers commonly used tools for deep learning and highlights applications in areas such as computer vision, natural language processing, and games. In the end, it discusses the future outlook and opportunities of deep learning.
Recurrent Neural Networks are popular Deep Learning models that have shown great promise to achieve state-of-the-art results in many tasks like Computer Vision, NLP, Finance and much more. Although being models proposed several years ago, RNN have gained popularity recently. In this talk, we will review how these models evolved over the years, dissection of RNN, current applications and its future.
This document provides an agenda for a presentation on deep learning, neural networks, convolutional neural networks, and interesting applications. The presentation will include introductions to deep learning and how it differs from traditional machine learning by learning feature representations from data. It will cover the history of neural networks and breakthroughs that enabled training of deeper models. Convolutional neural network architectures will be overviewed, including convolutional, pooling, and dense layers. Applications like recommendation systems, natural language processing, and computer vision will also be discussed. There will be a question and answer section.
This document summarizes Melanie Swan's presentation on deep learning. It began with defining key deep learning concepts and techniques, including neural networks, supervised vs. unsupervised learning, and convolutional neural networks. It then explained how deep learning works by using multiple processing layers to extract higher-level features from data and make predictions. Deep learning has various applications like image recognition and speech recognition. The presentation concluded by discussing how deep learning is inspired by concepts from physics and statistical mechanics.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session.
Artificial Neural Network seminar presentation using ppt.Mohd Faiz
- Artificial neural networks are inspired by biological neural networks and learning processes. They attempt to mimic the workings of the brain using simple units called artificial neurons that are connected in networks.
- Learning in neural networks involves modifying the synaptic strengths between neurons through mathematical optimization techniques. The goal is to minimize an error function that measures how well the network can approximate or complete a task.
- Neural networks can learn complex nonlinear functions through training algorithms like backpropagation that determine how to adjust the synaptic weights to improve performance on the learning task.
DLD meetup 2017, Efficient Deep LearningBrodmann17
The document discusses efficient techniques for deep learning on edge devices. It begins by noting that deep neural networks have high computational complexity which makes inference inefficient for edge devices without powerful GPUs. It then outlines the deep learning stack from hardware to libraries to frameworks to algorithms. The document focuses on how algorithms define model complexity and discusses the evolution of CNN architectures from LeNet5 to ResNet which generally increased in complexity. It covers techniques for reducing model size and operations like pruning, quantization, and knowledge distillation. The challenges of real-life applications on edge devices are discussed.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2019-embedded-vision-summit-wu
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Bichen Wu, Graduate Student Researcher in the EECS Department at the University of California, Berkeley, presents the "Enabling Automated Design of Computationally Efficient Deep Neural Networks" tutorial at the May 2019 Embedded Vision Summit.
Efficient deep neural networks are increasingly important in the age of AIoT (AI + IoT), in which people hope to deploy intelligent sensors and systems at scale. However, optimizing neural networks to achieve both high accuracy and efficient resource use on different target devices is difficult, since each device has its own idiosyncrasies.
In this talk, Wu introduces differentiable neural architecture search (DNAS), an approach for hardware-aware neural network architecture search. He shows that, using DNAS, the computation cost of the search itself is two orders of magnitude lower than previous approaches, while the models found by DNAS are optimized for target devices and surpass the previous state-of-the-art in efficiency and accuracy. Wu also explains how he used DNAS to find a new family of efficient neural networks called FBNets.
Architecture Design for Deep Neural Networks IWanjin Yu
This document summarizes Gao Huang's presentation on neural architectures for efficient inference. The presentation covered three parts: 1) macro-architecture innovations in convolutional neural networks (CNNs) such as ResNet, DenseNet, and multi-scale networks; 2) micro-architecture innovations including group convolution, depthwise separable convolution, and attention mechanisms; and 3) moving from static networks to dynamic networks that can adaptively select simpler or more complex models based on input complexity. The key idea is to enable faster yet accurate inference by matching computational cost to input difficulty.
In this talk, after a brief overview of AI concepts in particular Machine Learning (ML) techniques, some of the well-known computer design concepts for high performance and power efficiency are presented. Subsequently, those techniques that have had a promising impact for computing ML algorithms are discussed. Deep learning has emerged as a game changer for many applications in various fields of engineering and medical sciences. Although the primary computation function is matrix vector multiplication, many competing efficient implementations of this primary function have been proposed and put into practice. This talk will review and compare some of those techniques that are used for ML computer design.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2023/11/introduction-to-computer-vision-with-cnns-a-presentation-from-mohammad-haghighat/
Independent consultant Mohammad Haghighat presents the “Introduction to Computer Vision with Convolutional Neural Networks” tutorial at the May 2023 Embedded Vision Summit.
This presentation covers the basics of computer vision using convolutional neural networks. Haghighat begins by introducing some important conventional computer vision techniques and then transition to explaining the basics of machine learning and convolutional neural networks (CNNs) and showing how CNNs are used in visual perception.
Haghighat illustrates the building blocks and computational elements of neural networks through examples. This session provides an overview of how modern computer vision algorithms are designed, trained and used in real-world applications.
In this talk, an overview of current trends in machine learning will be discussed with an emphasize on challenges and opportunities facing this field. It will focus on deep learning methods and applications. Deep learning has emerged as one of the most promising research fields in artificial intelligence. The significant advancements that deep learning methods have brought about for large scale image classification tasks have generated a surge of excitement in applying the techniques to other problems in computer vision and more broadly into other disciplines of computer science. Moreover, the impact of machine learning on education, research, and economy will be briefly presented. The rapid growth of machine learning is positioned to impact our lives in a way that we have not been able to fully imagine. It behooves government leaders to take a lead in developing the necessary resources to ride the projected benefits of machine learning.
Opening Keynote at GTC 2015: Leaps in Visual ComputingNVIDIA
NVIDIA CEO and co-founder Jen-Hsun Huang took the stage for the GPU Technology Conference in the San Jose Convention Center to present some major announcements on March 17, 2015. You'll find out how NVIDIA is innovating in the field of deep learning, what NVIDIA DRIVE PX can do for automakers, and where Pascal, the next-generation GPU architecture, fits in the new performance roadmap.
Application of Neural Networks in Embedded Systems Applications (Ihor Starepr...IT Arena
Lviv IT Arena is a conference specially designed for programmers, designers, developers, top managers, inverstors, entrepreneurs and startuppers. Annually it takes place at the beginning of October in Lviv at Arena Lviv stadium. In 2016 the conference gathered more than 1800 participants and over 100 speakers from companies like Microsoft, Philips, Twitter, UBER and IBM. More details about the conference at itarena.lviv.ua.
This document discusses techniques for deploying deep learning models on low-power devices with limited compute resources. It describes methods such as parameter quantization, pruning parameters and connections, convolutional filter compression, matrix factorization, network architecture search, and knowledge distillation that can reduce model size and computational requirements while maintaining accuracy. Parameter quantization decreases model size by reducing precision. Pruning removes redundant connections and filters. Filter compression replaces large filters with smaller ones. Matrix factorization and architecture search optimize models for efficiency. Knowledge distillation transfers knowledge from large models to smaller, more efficient ones.
For the full video of this presentation, please visit:
http://www.embedded-vision.com/platinum-members/auvizsystems/embedded-vision-training/videos/pages/may-2015-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Nagesh Gupta, CEO and Founder of Auviz Systems, presents the "Trade-offs in Implementing Deep Neural Networks on FPGAs" tutorial at the May 2015 Embedded Vision Summit.
Video and images are a key part of Internet traffic—think of all the data generated by social networking sites such as Facebook and Instagram—and this trend continues to grow. Extracting usable information from video and images is thus a growing requirement in the data center. For example, object and face recognition are valuable for a wide range of uses, from social applications to security applications. Deep neural networks are currently the most popular form of convolutional neural networks (CNN) used in data centers for such applications. 3D convolutions are a core part of CNNs. Nagesh presents alternative implementations of 3D convolutions on FPGAs, and discusses trade-offs among them.
1. The document summarizes several papers on deep learning and convolutional neural networks. It discusses techniques like pruning weights, trained quantization, Huffman coding, and designing networks with fewer parameters like SqueezeNet.
2. One paper proposes techniques to compress deep neural networks by pruning, trained quantization, and Huffman coding to reduce model size. It evaluates these techniques on networks for MNIST and ImageNet, achieving compression rates of 35x to 49x with no loss of accuracy.
3. Another paper introduces SqueezeNet, a CNN architecture with AlexNet-level accuracy but 50x fewer parameters and a model size of less than 0.5MB. It employs fire modules with 1x1 convolutions to
This document discusses various techniques for optimizing deep neural network models and hardware for efficiency. It covers approaches such as exploiting activation and weight statistics, sparsity, compression, pruning neurons and synapses, decomposing trained filters, and knowledge distillation. The goal is to reduce operations, memory usage, and energy consumption to enable efficient inference on hardware like mobile phones and accelerators. Evaluation methodologies are also presented to guide energy-aware design space exploration.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/09/introduction-to-computer-vision-with-convolutional-neural-networks-a-presentation-from-ebay/
Mohammad Haghighat, Senior Manager for CoreAI at eBay, presents the “Introduction to Computer Vision with Convolutional Neural Networks” tutorial at the May 2024 Embedded Vision Summit.
This presentation covers the basics of computer vision using convolutional neural networks. Haghighat begins by introducing some important conventional computer vision techniques and then transitions to explaining the basics of machine learning and convolutional neural networks (CNNs) and showing how CNNs are used in visual perception.
Haghighat illustrates the building blocks and computational elements of neural networks through examples. You’ll gain a good overview of how modern computer vision algorithms are designed, trained and used in real-world applications.
Hardware Acceleration for Machine LearningCastLabKAIST
This document provides an overview of a lecture on hardware acceleration for machine learning. The lecture will cover deep neural network models like convolutional neural networks and recurrent neural networks. It will also discuss various hardware accelerators developed for machine learning, including those designed for mobile/edge and cloud computing environments. The instructor's background and the agenda topics are also outlined.
The next evolution in cloud computing is a smarter application not in the cloud. As the cloud has continued to evolve, the applications that utilize it have had more and more capabilities of the cloud. This presentation will show how to push logic and machine learning from the cloud to an edge application. Afterward, creating edge applications which utilize the intelligence of the cloud should become effortless.
AI gold rush, tool vendors and the next big thing
2017/12/27 at Mediatek
- Overview of booming AI applications, from media, entertainment, e-commerce, autonomous driving, surveillance, industrial inspection, medical imaging, bioinformatics, finance, etc., along with expert predictions of their market size and growth.
- Dissect the applications with largest size and growth into their technical components and their unmet demands.
- Among all the unmet demands and uncertainties in this AI gold rush, what should an IC design company do? I’ll briefly cover NVIDIA’s case, which most of us know well already, then supplement case studies of Qualcomm, Intel, Google TPU and other smaller firms.
Even when we have a clear target, it takes years for supporting libraries and software to be properly optimized. I’ll share some thoughts and personal experiences on how to make sequentially-ordered hardware/software/library optimization happen faster and in parallel, and the tools that the IC design house need to provide in order for it to happen.
Deep learning and applications in non-cognitive domains IDeakin University
This document outlines an agenda for a presentation on deep learning and its applications in non-cognitive domains. The presentation is divided into three parts: an introduction to deep learning theory, applying deep learning to non-cognitive domains in practice, and advanced topics. The introduction covers neural network architectures like feedforward, recurrent, and convolutional networks. It also discusses techniques for improving training like rectified linear units and skip connections. The practice section will provide hands-on examples in domains like healthcare and software engineering. The advanced topics section will discuss unsupervised learning, structured outputs, and positioning techniques in deep learning.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/11/improved-data-sampling-techniques-for-training-neural-networks-a-presentation-from-karthik-rao-aroor/
Independent AI Engineer Karthik Rao Aroor presents the “Improved Data Sampling Techniques for Training Neural Networks” tutorial at the May 2024 Embedded Vision Summit.
For classification problems in which there are equal numbers of samples in each class, Aroor proposes and presents a novel mini-batch sampling approach to train neural networks using gradient descent. His proposed approach ensures a uniform distribution of samples from all classes in a mini-batch. He shares results showing that this approach yields faster convergence than the random sampling approach commonly used today.
Aroor illustrates his approach using several neural network models trained on commonly used datasets, including a truncated version of ImageNet. He also presents results for large and small mini-batch sizes relative to the number of classes. Comparing these results to a suboptimal sampling approach, he hypothesizes that having a uniform distribution of samples from each class in a mini-batch is an optimal sampling approach. His approach benefits model trainers by achieving higher model accuracy with reduced training time.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/cost-efficient-high-quality-ai-for-consumer-grade-smart-home-cameras-a-presentation-from-wyze/
Lin Chen, Chief Scientist at Wyze, presents the “Cost-efficient, High-quality AI for Consumer-grade Smart Home Cameras” tutorial at the May 2024 Embedded Vision Summit.
In this talk, Chen explains how Wyze delivers robust visual AI at ultra-low cost for millions of consumer smart cameras, and how his company is rapidly expanding these capabilities. He begins by introducing Wyze’s edge-AI-enabled consumer cameras and related cloud services, which were created based on goals of delivering “too good to be true” value and building long-term relationships with customers.
Chen explains how Wyze has improved model accuracy with user feedback while keeping hardware resource usage to a minimum. Next, he shows how the company has expanded its cameras’ AI capabilities to include things like pet, vehicle and package detection. He also shows how Wyze enables customers to utilize smart cameras as “supersensors” for home security, replacing conventional single-purpose sensors and overcoming many of their limitations. Finally, he shows how the company is empowering users to add their own personalized AI capabilities, enabling support for numerous diverse use cases.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/edge-ai-optimization-on-rails-literally-a-presentation-from-wabtec/
Matthew Pietrzykowski, Principal Data Scientist at Wabtec, presents the “Edge AI Optimization on Rails—Literally” tutorial at the May 2024 Embedded Vision Summit.
In this talk, Pietrzykowski shares highlights from his company’s adventures developing computer vision solutions for the rail transportation industry. He begins with an introduction to the types of machine perception problems encountered in this unique industry. He then shares insights gained during his company's implementation of a perception system for rail auditing.
In particular, Pietrzykowski discusses the challenges faced in implementing multiple optimized CNNs in a constrained compute system, and how his company addressed these challenges. He also explains trade-offs associated with using classical homographic techniques in combination with neural networks.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/how-large-language-models-are-impacting-computer-vision-a-presentation-from-voxel51/
Jacob Marks, Senior ML Engineer and Researcher at Voxel51, presents the “How Large Language Models Are Impacting Computer Vision” tutorial at the May 2024 Embedded Vision Summit.
Large language models (LLMs) are revolutionizing the way we interact with computers and the world around us. However, in order to truly understand the world, LLM-powered agents need to be able to see.
Will models in production be multimodal, or will text-only LLMs leverage purpose-built vision models as tools? Where do techniques like multimodal retrieval-augmented generation (RAG) fit in? In this talk, Marks gives an overview of key LLM-centered projects that are reshaping the field of computer vision and discusses where we are headed in a multimodal world.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/implementing-ai-computer-vision-for-corporate-security-surveillance-a-presentation-from-vmware/
Prasad Saranjame, former Head of Physical Security and Resiliency at VMware, presents the “Implementing AI/Computer Vision for Corporate Security Surveillance” tutorial at the May 2024 Embedded Vision Summit.
AI-enabled security cameras offer substantial benefits for corporate security and operational efficiency. However, successful deployment requires thoughtful selection of use cases and meticulous attention to managing the technology’s impact on teams. This talk explores VMware’s journey through a multiyear digital transformation centered on AI/ML-based security monitoring, decision-making and response.
Saranjame discusses the business, technology and organizational challenges faced, along with best practices for overcoming them. Key strategies include running proof of concept projects, focusing on operational excellence to drive cost reductions and investing in change management to ensure team buy-in. Additionally, he provides insights for others looking to deploy AI in their operations, emphasizing disciplined prioritization, strategic partnerships and a focus on long-term scalability over short-term cost reduction. You will gain actionable insights for effectively harnessing AI in your own organizations.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/continual-learning-thru-sequential-lightweight-optimization-a-presentation-from-vision-elements/
Guy Lavi, Managing Partner at Vision Elements, presents the “Continual, On-the-fly Learning through Sequential, Lightweight Optimization” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, Lavi shows how techniques of sequential optimization are applied to enable continual learning during run-time, as new observations flow in. The lightweight nature of these techniques, using only the new batches of observations for processing, allows for new training iterations to be performed on the edge without losing memory of the entire pool of observations used for the initial training. He presents detailed examples using this technique, showing how it can be used to optimize a linear function, an image warping algorithm and an object classification neural network.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/multi-object-tracking-systems-a-presentation-from-tryolabs/
Javier Berneche, Senior Machine Learning Engineer at Tryolabs, presents the “Multiple Object Tracking Systems” tutorial at the May 2024 Embedded Vision Summit.
Object tracking is an essential capability in many computer vision systems, including applications in fields such as traffic control, self-driving vehicles, sports and more. In this talk, Berneche walks through the construction of a typical multiple object tracking (MOT) algorithm step by step. At each step, he identifies key challenges and explores design choices (for example, detection-based vs. detection-free approaches and online vs. offline tracking).
Berneche discusses available off-the-shelf MOT algorithms and open-source libraries. He also identifies areas where current MOT algorithms fall short. And he introduces metrics and benchmarks commonly used to evaluate MOT solutions.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/improved-navigation-assistance-for-the-blind-via-real-time-edge-ai-a-presentation-from-tesla/
Aishwarya Jadhav, Software Engineer in the Autopilot AI Team at Tesla, presents the “Improved Navigation Assistance for the Blind via Real-time Edge AI,” tutorial at the May 2024 Embedded Vision Summit.
In this talk, Jadhav presents recent work on AI Guide Dog, a groundbreaking research project aimed at providing navigation assistance for the blind community. This multiyear project at Carnegie Mellon University leverages AI to predict sighted human reactions in real time and convey this information audibly to blind individuals, overcoming the limitations of existing GPS apps and mobility tools for the blind.
Jadhav discusses the various vision-only and multimodal models evaluated. She also discusses imitation learning approaches currently being explored. In addition, she highlights trade-offs among the strict requirements for models to ensure explainable predictions, high accuracy and real-time processing on mobile devices. And she shares insights gained through three iterations of this project, explaining data collection procedures, training pipelines and cutting-edge vision and multimodal modeling methodologies. She concludes with some exciting results.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/using-vision-systems-generative-models-and-reinforcement-learning-for-sports-analytics-a-presentation-from-sportlogiq/
Mehrsan Javan, Chief Technology Officer at Sportlogiq, presents the “Using Vision Systems, Generative Models and Reinforcement Learning for Sports Analytics” tutorial at the May 2024 Embedded Vision Summit.
At a high level, sport analytics systems can be broken into two components: sensory data collection and analytical models that turn sensory data into insights for users. In this talk, Javan focuses on the latter, and more specifically on the challenges his company has encountered in adapting advanced analytics originally developed for professional leagues to create a new product for use in a new market—youth sports.
These challenges arise due to the unfamiliarity of end users with sophisticated analytical metrics, incomplete and partially accurate underlying visual data and the inherent limitations of vision-based data collection systems. Javan explains how Sportlogiq uses a combination of vision systems, generative models and reinforcement learning techniques to develop compelling products for youth sports, and shares what’s been learned in this process.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/introduction-to-cameras-for-embedded-applications-a-presentation-from-sensorspace/
Brian Rodricks, CTO of SensorSpace, presents the “Introduction to Cameras for Embedded Applications” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, Rodricks introduces the essential features of cameras for embedded vision applications. He explains lens mounts and camera interface options such as MIPI, USB and GigE. He explores sensor features and trade-offs, including rolling shutter vs. global shutter, monochrome vs. color, RGB vs. RGBW and pixel size vs. array size.
Rodricks also examines trade-offs between field of view and distortion, as well as performance trade-offs such as dynamic range vs. sensitivity, pixel size vs. resolution and frame rates. Finally, he addresses the crucial aspects of price vs. performance and control, covering access to register settings, raw data and software.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/introduction-to-modern-radar-for-machine-perception-a-presentation-from-sensor-cortek/
Robert Laganière, Professor at the University of Ottawa and CEO of Sensor Cortek, presents the “Introduction to Modern Radar for Machine Perception” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, Laganière provides an introduction to radar (short for radio detection and ranging) for machine perception. Radar is a proven technology with a long history of successful development and it plays an increasingly important role in the deployment of robust perception systems.
Laganière explains how radar sensors work—in particular, how radio waves are used to accomplish detection and ranging. He explains key concepts behind this technology, including Doppler effect, time-of-flight, frequency modulation and continuous waves (FMCW). Finally, he explores the main advantages and disadvantages of radar for machine perception.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/diagnosing-problems-and-implementing-solutions-for-deep-neural-network-training-a-presentation-from-sensor-cortek/
Fahed Hassanat, COO and Head of Engineering at Sensor Cortek, presents the “Deep Neural Network Training: Diagnosing Problems and Implementing Solutions” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, Hassanat delves into some of the most common problems that arise when training deep neural networks. He provides a brief overview of essential training metrics, including accuracy, precision, false positives, false negatives and F1 score.
Hassanat then explores training challenges that arise from problems with hyperparameters, inappropriately sized models, inadequate models, poor-quality datasets, imbalances within training datasets and mismatches between training and testing datasets. To help detect and diagnose training problems, he covers techniques such as understanding performance curves, recognizing overfitting and underfitting, analyzing confusion matrices and identifying class interaction issues.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/seeing-through-machines-a-guide-to-image-sensors-for-edge-ai-applications-a-presentation-from-seedar-consulting/
Armita Abadian, Advisor to SEEdar Consulting, presents the “Seeing Through Machines: A Guide to Image Sensors for Edge AI Applications” tutorial at the May 2024 Embedded Vision Summit.
Imagine a robot navigating a busy factory floor, a self-driving car detecting obstacles on the road or a security system accurately recognizing faces. These advancements in artificial intelligence are powered by tiny devices called CMOS image sensors. With a vast array of sensor options available, choosing the right one for your specific application can be challenging.
In this talk, Abadian demystifies the world of image sensors, exploring how they work and uncovering the key performance criteria you need to consider for projects in automation, robotics and beyond. You’ll be able to navigate the sensor landscape, equipped with the knowledge to make informed choices and unlock the potential of edge AI.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/transformer-networks-how-they-work-and-why-they-matter-a-presentation-from-ryddle-ai/
Rakshit Agrawal, Co-Founder and CEO of Ryddle AI, presents the “Transformer Networks: How They Work and Why They Matter” tutorial at the May 2024 Embedded Vision Summit.
Transformer neural networks have revolutionized artificial intelligence by introducing an architecture built around self-attention mechanisms. This has enabled unprecedented advances in understanding sequential data, such as human languages, while also dramatically improving accuracy on nonsequential tasks like object detection.
In this talk, Agrawal explains the technical underpinnings of transformer architectures, with particular focus on self-attention mechanisms. He also explores how transformers have influenced the direction of AI research and industry innovation. Finally, he touches on ethical considerations and discusses how transformers are likely to evolve in the near future.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/removing-weather-related-image-degradation-at-the-edge-a-presentation-from-rivian/
Ramit Pahwa, Machine Learning Scientist at Rivian, presents the “Removing Weather-related Image Degradation at the Edge” tutorial at the May 2024 Embedded Vision Summit.
For machines that operate outdoors—such as autonomous cars and trucks—image quality degradation due to weather conditions presents a significant challenge. For example, snow, rainfall and raindrops on optical surfaces can wreak havoc on machine perception algorithms. In this talk, Pahwa explains the key challenges in restoring images degraded by weather, such as lack of annotated datasets, and the need for multiple models to address different types of image degradation.
Pahwa also introduces metrics for assessing image degradation. He then explains Rivian’s solutions and shares results, demonstrating the efficacy of transformer-based models and of a novel, language-driven, all-in-one model for image restoration. Finally, he highlights the techniques used to create efficient implementations of Rivian’s models for deployment at the edge—including quantization and pruning—and shares lessons learned from implementing these models on a target processor.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/seeing-the-invisible-unveiling-hidden-details-through-advanced-image-acquisition-techniques-a-presentation-from-qualitas-technologies/
Raghava Kashyapa, CEO of Qualitas Technologies, presents the “Seeing the Invisible: Unveiling Hidden Details through Advanced Image Acquisition Techniques” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, Kashyapa explores how advanced image acquisition techniques reveal previously unseen information, improving the ability of algorithms to provide valuable insights. He introduces various techniques, including high-resolution optics, filters, high dynamic range imaging and multispectral imaging.
Kashyapa explains how these techniques work and shares recommendations on how system developers can best utilize them to improve image quality and boost the performance of innovative machine vision applications. You’ll learn and explore how to make the invisible visible in your projects.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/data-efficient-and-generalizable-the-domain-specific-small-vision-model-revolution-a-presentation-from-pixel-scientia-labs/
Heather Couture, Founder and Computer Vision Consultant at Pixel Scientia Labs, presents the “Data-efficient and Generalizable: The Domain-specific Small Vision Model Revolution” tutorial at the May 2024 Embedded Vision Summit.
Large vision models (LVMs) trained on a large and diverse set of imagery are revitalizing computer vision, just as LLMs did for language modeling. However, LVMs are not nearly as effective when applied to unique types of imagery. To handle labeled data scarcity without overfitting, we need models that are tuned to a specific domain of imagery. Whether it’s a single medical imaging modality, multispectral drone photos or snapshots from a manufacturing line, these fine-grained applications are best captured with a model that can accommodate the available data.
A small vision model with fewer parameters improves generalizability with the added bonus of better computational efficiency so that it can run on an edge device. In this talk, Couture shows why domain-specific models are essential and how they can be trained without labeled data. She concludes by demonstrating the efficacy of domain-specific models in handling small training sets, imbalanced data and distribution shifts for various types of imagery.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/omnilert-gun-detect-harnessing-computer-vision-to-tackle-gun-violence-a-presentation-from-omnilert/
Chad Green, Director of Artificial Intelligence at Omnilert, presents the “Omnilert Gun Detect: Harnessing Computer Vision to Tackle Gun Violence” tutorial at the May 2024 Embedded Vision Summit.
In the United States in 2023, there were 658 mass shootings, and 42,996 people lost their lives to gun violence. Detecting and rapidly responding to potential and actual shootings in an automated fashion is critical to reducing these tragic figures. In 2020, Omnilert, a pioneer in emergency notification systems, launched Omnilert Gun Detect, an AI-powered platform that combines gun detection, verification, activation of security systems and notification.
In this talk, Green describes the development of Omnilert Gun Detect. He covers why computer vision is the right solution to this problem, how Omnilert went about building the product and the business and technical challenges overcome along the way. He also talks about Omnilert’s market traction, and concludes with lessons learned in building this important system.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/adventures-in-moving-a-computer-vision-solution-from-cloud-to-edge-a-presentation-from-metaconsumer/
Nate D’Amico, CTO and Head of Product at MetaConsumer, presents the “Adventures in Moving a Computer Vision Solution from Cloud to Edge” tutorial at the May 2024 Embedded Vision Summit.
Optix is a computer vision-based AI system that measures advertising and media exposures on mobile devices for real-time marketing optimization. Optix was initially developed as a cloud-based solution, but costs and limitations associated with relying entirely on the cloud drove MetaConsumer to implement an edge solution.
In this talk, D’Amico introduces his company’s application, the role that computer vision plays in it and the challenges of deploying it at scale. He shares the lessons learned from operating a cloud-based solution at scale, the trade-offs that drove MetaConsumer to create an edge-based solution and the hardware and software challenges faced in implementing it.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/10/bridging-vision-and-language-designing-training-and-deploying-multimodal-large-language-models-a-presentation-from-meta-reality-labs/
Adel Ahmadyan, Staff Engineer at Meta Reality Labs, presents the “Bridging Vision and Language: Designing, Training and Deploying Multimodal Large Language Models” tutorial at the May 2024 Embedded Vision Summit.
In this talk, Ahmadyan explores the use of multimodal large language models in real-world edge applications. He begins by explaining how these large multimodal models (LMMs) work and highlighting their key components, giving special attention to how LMMs merge understanding in the vision and language domains.
Next, Ahmadyan discusses the process of training LMMs and the types of data needed to tune them for specific tasks. Finally, he highlights some of the key challenges in deploying LMMs in resource-constrained edge devices and shares techniques for overcoming these challenges.
DePIN = Real-World Infra + Blockchain
DePIN stands for Decentralized Physical Infrastructure Networks.
It connects physical devices to Web3 using token incentives.
How Does It Work?
Individuals contribute to infrastructure like:
Wireless networks (e.g., Helium)
Storage (e.g., Filecoin)
Sensors, compute, and energy
They earn tokens for their participation.
Agentic AI - The New Era of IntelligenceMuzammil Shah
This presentation is specifically designed to introduce final-year university students to the foundational principles of Agentic Artificial Intelligence (AI). It aims to provide a clear understanding of how Agentic AI systems function, their key components, and the underlying technologies that empower them. By exploring real-world applications and emerging trends, the session will equip students with essential knowledge to engage with this rapidly evolving area of AI, preparing them for further study or professional work in the field.
As data privacy regulations become more pervasive across the globe and organizations increasingly handle and transfer (including across borders) meaningful volumes of personal and confidential information, the need for robust contracts to be in place is more important than ever.
This webinar will provide a deep dive into privacy contracting, covering essential terms and concepts, negotiation strategies, and key practices for managing data privacy risks.
Whether you're in legal, privacy, security, compliance, GRC, procurement, or otherwise, this session will include actionable insights and practical strategies to help you enhance your agreements, reduce risk, and enable your business to move fast while protecting itself.
This webinar will review key aspects and considerations in privacy contracting, including:
- Data processing addenda, cross-border transfer terms including EU Model Clauses/Standard Contractual Clauses, etc.
- Certain legally-required provisions (as well as how to ensure compliance with those provisions)
- Negotiation tactics and common issues
- Recent lessons from recent regulatory actions and disputes
AI in Java - MCP in Action, Langchain4J-CDI, SmallRye-LLM, Spring AIBuhake Sindi
This is the presentation I gave with regards to AI in Java, and the work that I have been working on. I've showcased Model Context Protocol (MCP) in Java, creating server-side MCP server in Java. I've also introduced Langchain4J-CDI, previously known as SmallRye-LLM, a CDI managed too to inject AI services in enterprise Java applications. Also, honourable mention: Spring AI.
UiPath Community Berlin: Studio Tips & Tricks and UiPath InsightsUiPathCommunity
Join the UiPath Community Berlin (Virtual) meetup on May 27 to discover handy Studio Tips & Tricks and get introduced to UiPath Insights. Learn how to boost your development workflow, improve efficiency, and gain visibility into your automation performance.
📕 Agenda:
- Welcome & Introductions
- UiPath Studio Tips & Tricks for Efficient Development
- Best Practices for Workflow Design
- Introduction to UiPath Insights
- Creating Dashboards & Tracking KPIs (Demo)
- Q&A and Open Discussion
Perfect for developers, analysts, and automation enthusiasts!
This session streamed live on May 27, 18:00 CET.
Check out all our upcoming UiPath Community sessions at:
👉 https://community.uipath.com/events/
Join our UiPath Community Berlin chapter:
👉 https://community.uipath.com/berlin/
Content and eLearning Standards: Finding the Best Fit for Your-TrainingRustici Software
Tammy Rutherford, Managing Director of Rustici Software, walks through the pros and cons of different standards to better understand which standard is best for your content and chosen technologies.
Dev Dives: System-to-system integration with UiPath API WorkflowsUiPathCommunity
Join the next Dev Dives webinar on May 29 for a first contact with UiPath API Workflows, a powerful tool purpose-fit for API integration and data manipulation!
This session will guide you through the technical aspects of automating communication between applications, systems and data sources using API workflows.
📕 We'll delve into:
- How this feature delivers API integration as a first-party concept of the UiPath Platform.
- How to design, implement, and debug API workflows to integrate with your existing systems seamlessly and securely.
- How to optimize your API integrations with runtime built for speed and scalability.
This session is ideal for developers looking to solve API integration use cases with the power of the UiPath Platform.
👨🏫 Speakers:
Gunter De Souter, Sr. Director, Product Manager @UiPath
Ramsay Grove, Product Manager @UiPath
This session streamed live on May 29, 2025, 16:00 CET.
Check out all our upcoming UiPath Dev Dives sessions:
👉 https://community.uipath.com/dev-dives-automation-developer-2025/
What’s New in Web3 Development Trends to Watch in 2025.pptxLisa ward
Emerging Web3 development trends in 2025 include AI integration, enhanced scalability, decentralized identity, and increased enterprise adoption of blockchain technologies.
Unlock your organization’s full potential with the 2025 Digital Adoption Blueprint. Discover proven strategies to streamline software onboarding, boost productivity, and drive enterprise-wide digital transformation.
For those who have ever wanted to recreate classic games, this presentation covers my five-year journey to build a NES emulator in Kotlin. Starting from scratch in 2020 (you can probably guess why), I’ll share the challenges posed by the architecture of old hardware, performance optimization (surprise, surprise), and the difficulties of emulating sound. I’ll also highlight which Kotlin features shine (and why concurrency isn’t one of them). This high-level overview will walk through each step of the process—from reading ROM formats to where GPT can help, though it won’t write the code for us just yet. We’ll wrap up by launching Mario on the emulator (hopefully without a call from Nintendo).
Master tester AI toolbox - Kari Kakkonen at Testaus ja AI 2025 ProfessioKari Kakkonen
My slides at Professio Testaus ja AI 2025 seminar in Espoo, Finland.
Deck in English, even though I talked in Finnish this time, in addition to chairing the event.
I discuss the different motivations for testing to use AI tools to help in testing, and give several examples in each categories, some open source, some commercial.
"AI in the browser: predicting user actions in real time with TensorflowJS", ...Fwdays
With AI becoming increasingly present in our everyday lives, the latest advancements in the field now make it easier than ever to integrate it into our software projects. In this session, we’ll explore how machine learning models can be embedded directly into front-end applications. We'll walk through practical examples, including running basic models such as linear regression and random forest classifiers, all within the browser environment.
Once we grasp the fundamentals of running ML models on the client side, we’ll dive into real-world use cases for web applications—ranging from real-time data classification and interpolation to object tracking in the browser. We'll also introduce a novel approach: dynamically optimizing web applications by predicting user behavior in real time using a machine learning model. This opens the door to smarter, more adaptive user experiences and can significantly improve both performance and engagement.
In addition to the technical insights, we’ll also touch on best practices, potential challenges, and the tools that make browser-based machine learning development more accessible. Whether you're a developer looking to experiment with ML or someone aiming to bring more intelligence into your web apps, this session will offer practical takeaways and inspiration for your next project.