See hints, Ref under each slide
Deep Learning tutorial
https://www.youtube.com/watch?v=q4rZ9ujp3bw&list=PLAI6JViu7XmflH_eGgsWkwvv6lbXhYjjY
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session.
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...Simplilearn
This Deep Learning Presentation will help you in understanding what is Deep learning, why do we need Deep learning, applications of Deep Learning along with a detailed explanation on Neural Networks and how these Neural Networks work. Deep learning is inspired by the integral function of the human brain specific to artificial neural networks. These networks, which represent the decision-making process of the brain, use complex algorithms that process data in a non-linear way, learning in an unsupervised manner to make choices based on the input. This Deep Learning tutorial is ideal for professionals with beginners to intermediate levels of experience. Now, let us dive deep into this topic and understand what Deep learning actually is.
Below topics are explained in this Deep Learning Presentation:
1. What is Deep Learning?
2. Why do we need Deep Learning?
3. Applications of Deep Learning
4. What is Neural Network?
5. Activation Functions
6. Working of Neural Network
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change. With this Tensorflow course, you’ll build expertise in deep learning models, learn to operate TensorFlow to manage neural networks and interpret the results.
You can gain in-depth knowledge of Deep Learning by taking our Deep Learning certification training course. With Simplilearn’s Deep Learning course, you will prepare for a career as a Deep Learning engineer as you master concepts and techniques including supervised and unsupervised learning, mathematical and heuristic aspects, and hands-on modeling to develop algorithms.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
It’s long ago, approx. 30 years, since AI was not only a topic for Science-Fiction writers, but also a major research field surrounded with huge hopes and investments. But the over-inflated expectations ended in a subsequent crash and followed by a period of absent funding and interest – the so-called AI winter. However, the last 3 years changed everything – again. Deep learning, a machine learning technique inspired by the human brain, successfully crushed one benchmark after another and tech companies, like Google, Facebook and Microsoft, started to invest billions in AI research. “The pace of progress in artificial general intelligence is incredible fast” (Elon Musk – CEO Tesla & SpaceX) leading to an AI that “would be either the best or the worst thing ever to happen to humanity” (Stephen Hawking – Physicist).
What sparked this new Hype? How is Deep Learning different from previous approaches? Are the advancing AI technologies really a threat for humanity? Let’s look behind the curtain and unravel the reality. This talk will explore why Sundar Pichai (CEO Google) recently announced that “machine learning is a core transformative way by which Google is rethinking everything they are doing” and explain why "Deep Learning is probably one of the most exciting things that is happening in the computer industry” (Jen-Hsun Huang – CEO NVIDIA).
Either a new AI “winter is coming” (Ned Stark – House Stark) or this new wave of innovation might turn out as the “last invention humans ever need to make” (Nick Bostrom – AI Philosoph). Or maybe it’s just another great technology helping humans to achieve more.
The document discusses power flow analysis, which determines voltages, currents, real power, and reactive power in a power system under steady-state load conditions. It describes the different types of buses in a power system and how they are modeled. The key component of power flow is the bus admittance matrix, which relates nodal voltages to branch currents based on Kirchhoff's current law. Solving the matrix equations provides the voltage magnitude and angle at each bus.
Mathematics of the number 369 and the power of universal resistance.pdfWim van Es
This booklet describes the number 369 introduced by Nikola Tesla. The numbers 369 are connected to a new force of nature. Resistance force. In addition, this booklet describes two new unique golden spirals based on the Golden Pyramid. And it describes a new view of comets.
What is an "ensemble learner"? How can we combine different base learners into an ensemble in order to improve the overall classification performance? In this lecture, we are providing some answers to these questions.
The document provides an overview of Long Short Term Memory (LSTM) networks. It discusses:
1) The vanishing gradient problem in traditional RNNs and how LSTMs address it through gated cells that allow information to persist without decay.
2) The key components of LSTMs - forget gates, input gates, output gates and cell states - and how they control the flow of information.
3) Common variations of LSTMs including peephole connections, coupled forget/input gates, and Gated Recurrent Units (GRUs). Applications of LSTMs in areas like speech recognition, machine translation and more are also mentioned.
The document provides an overview of deep learning, including its past, present, and future. It discusses the concepts of artificial general intelligence, artificial superintelligence, and predictions about their development from experts like Hawking, Musk, and Gates. Key deep learning topics are summarized, such as neural networks, machine learning approaches, important algorithms and researchers, and how deep learning works.
Introduction to Recurrent Neural NetworkKnoldus Inc.
The document provides an introduction to recurrent neural networks (RNNs). It discusses how RNNs differ from feedforward neural networks in that they have internal memory and can use their output from the previous time step as input. This allows RNNs to process sequential data like time series. The document outlines some common RNN types and explains the vanishing gradient problem that can occur in RNNs due to multiplication of small gradient values over many time steps. It discusses solutions to this problem like LSTMs and techniques like weight initialization and gradient clipping.
Introduction For seq2seq(sequence to sequence) and RNNHye-min Ahn
This is my slides for introducing sequence to sequence model and Recurrent Neural Network(RNN) to my laboratory colleagues.
Hyemin Ahn, @CPSLAB, Seoul National University (SNU)
Deep Learning Tutorial | Deep Learning TensorFlow | Deep Learning With Neural...Simplilearn
This Deep Learning presentation will help you in understanding what is Deep Learning, why do we need Deep learning, what is neural network, applications of Deep Learning, what is perceptron, implementing logic gates using perceptron, types of neural networks. At the end of the video, you will get introduced to TensorFlow along with a usecase implementation on recognizing hand-written digits. Deep Learning is inspired by the integral function of the human brain specific to artificial neural networks. These networks, which represent the decision-making process of the brain, use complex algorithms that process data in a non-linear way, learning in an unsupervised manner to make choices based on the input. Deep Learning, on the other hand, uses advanced computing power and special type of neural networks and applies them to large amounts of data to learn, understand, and identify complicated patterns. W will also understand neural networks and how they work in this Deep Learning tutorial video. This Deep Learning tutorial is ideal for professionals with beginner to intermediate level of experience. Now, let us dive deep into this topic and understand what Deep Learning actually is.
Below topics are explained in this Deep Learning presentation:
1. What is Deep Learning?
2. Why do we need Deep Learning?
3. What is Neural network?
4. What is Perceptron?
5. Implementing logic gates using Perceptron
6. Types of Neural networks
7. Applications of Deep Learning
8. Working of Neural network
9. Introduction to TensorFlow
10. Use case implementation using TensorFlow
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
Part 2 of the Deep Learning Fundamentals Series, this session discusses Tuning Training (including hyperparameters, overfitting/underfitting), Training Algorithms (including different learning rates, backpropagation), Optimization (including stochastic gradient descent, momentum, Nesterov Accelerated Gradient, RMSprop, Adaptive algorithms - Adam, Adadelta, etc.), and a primer on Convolutional Neural Networks. The demos included in these slides are running on Keras with TensorFlow backend on Databricks.
Our fall 12-Week Data Science bootcamp starts on Sept 21st,2015. Apply now to get a spot!
If you are hiring Data Scientists, call us at (1)888-752-7585 or reach [email protected] to share your openings and set up interviews with our excellent students.
---------------------------------------------------------------
Come join our meet-up and learn how easily you can use R for advanced Machine learning. In this meet-up, we will demonstrate how to understand and use Xgboost for Kaggle competition. Tong is in Canada and will do remote session with us through google hangout.
---------------------------------------------------------------
Speaker Bio:
Tong is a data scientist in Supstat Inc and also a master students of Data Mining. He has been an active R programmer and developer for 5 years. He is the author of the R package of XGBoost, one of the most popular and contest-winning tools on kaggle.com nowadays.
Pre-requisite(if any): R /Calculus
Preparation: A laptop with R installed. Windows users might need to have RTools installed as well.
Agenda:
Introduction of Xgboost
Real World Application
Model Specification
Parameter Introduction
Advanced Features
Kaggle Winning Solution
Event arrangement:
6:45pm Doors open. Come early to network, grab a beer and settle in.
7:00-9:00pm XgBoost Demo
Reference:
https://github.com/dmlc/xgboost
This document provides an agenda for a presentation on deep learning, neural networks, convolutional neural networks, and interesting applications. The presentation will include introductions to deep learning and how it differs from traditional machine learning by learning feature representations from data. It will cover the history of neural networks and breakthroughs that enabled training of deeper models. Convolutional neural network architectures will be overviewed, including convolutional, pooling, and dense layers. Applications like recommendation systems, natural language processing, and computer vision will also be discussed. There will be a question and answer section.
Part 1 of the Deep Learning Fundamentals Series, this session discusses the use cases and scenarios surrounding Deep Learning and AI; reviews the fundamentals of artificial neural networks (ANNs) and perceptrons; discuss the basics around optimization beginning with the cost function, gradient descent, and backpropagation; and activation functions (including Sigmoid, TanH, and ReLU). The demos included in these slides are running on Keras with TensorFlow backend on Databricks.
Basic concept of Deep Learning with explaining its structure and backpropagation method and understanding autograd in PyTorch. (+ Data parallism in PyTorch)
Recurrent Neural Networks have shown to be very powerful models as they can propagate context over several time steps. Due to this they can be applied effectively for addressing several problems in Natural Language Processing, such as Language Modelling, Tagging problems, Speech Recognition etc. In this presentation we introduce the basic RNN model and discuss the vanishing gradient problem. We describe LSTM (Long Short Term Memory) and Gated Recurrent Units (GRU). We also discuss Bidirectional RNN with an example. RNN architectures can be considered as deep learning systems where the number of time steps can be considered as the depth of the network. It is also possible to build the RNN with multiple hidden layers, each having recurrent connections from the previous time steps that represent the abstraction both in time and space.
- Naive Bayes is a classification technique based on Bayes' theorem that uses "naive" independence assumptions. It is easy to build and can perform well even with large datasets.
- It works by calculating the posterior probability for each class given predictor values using the Bayes theorem and independence assumptions between predictors. The class with the highest posterior probability is predicted.
- It is commonly used for text classification, spam filtering, and sentiment analysis due to its fast performance and high success rates compared to other algorithms.
The document discusses Long Short Term Memory (LSTM) networks, which are a type of recurrent neural network capable of learning long-term dependencies. It explains that unlike standard RNNs, LSTMs use forget, input, and output gates to control the flow of information into and out of the cell state, allowing them to better capture long-range temporal dependencies in sequential data like text, audio, and time-series data. The document provides details on how LSTM gates work and how LSTMs can be used for applications involving sequential data like machine translation and question answering.
1) Deep learning is a type of machine learning that uses neural networks with many layers to learn representations of data with multiple levels of abstraction.
2) Deep learning techniques include unsupervised pretrained networks, convolutional neural networks, recurrent neural networks, and recursive neural networks.
3) The advantages of deep learning include automatic feature extraction from raw data with minimal human effort, and surpassing conventional machine learning algorithms in accuracy across many data types.
Tijmen Blankenvoort, co-founder Scyfer BV, presentation at Artificial Intelligence Meetup 15-1-2014. Introduction into Neural Networks and Deep Learning.
Artificial Intelligence, Machine Learning, Deep Learning
The 5 myths of AI
Deep Learning in action
Basics of Deep Learning
NVIDIA Volta V100 and AWS P3
Machine learning helps predict behavior and recognize patterns that humans cannot by learning from data without relying on programmed rules. It is an algorithmic approach that differs from statistical modeling which formalizes relationships through mathematical equations. Machine learning is a part of the broader field of artificial intelligence which aims to develop systems that can act and respond intelligently like humans. The machine learning workflow involves collecting and preprocessing data, selecting algorithms, training models, and evaluating performance. Common machine learning algorithms include supervised learning, unsupervised learning, reinforcement learning, and deep learning. Popular tools for machine learning include Python, R, TensorFlow, and Spark.
The document provides an overview of convolutional neural networks (CNNs) and their layers. It begins with an introduction to CNNs, noting they are a type of neural network designed to process 2D inputs like images. It then discusses the typical CNN architecture of convolutional layers followed by pooling and fully connected layers. The document explains how CNNs work using a simple example of classifying handwritten X and O characters. It provides details on the different layer types, including convolutional layers which identify patterns using small filters, and pooling layers which downsample the inputs.
Deep learning is introduced along with its applications and key players in the field. The document discusses the problem space of inputs and outputs for deep learning systems. It describes what deep learning is, providing definitions and explaining the rise of neural networks. Key deep learning architectures like convolutional neural networks are overviewed along with a brief history and motivations for deep learning.
The document provides an overview of Long Short Term Memory (LSTM) networks. It discusses:
1) The vanishing gradient problem in traditional RNNs and how LSTMs address it through gated cells that allow information to persist without decay.
2) The key components of LSTMs - forget gates, input gates, output gates and cell states - and how they control the flow of information.
3) Common variations of LSTMs including peephole connections, coupled forget/input gates, and Gated Recurrent Units (GRUs). Applications of LSTMs in areas like speech recognition, machine translation and more are also mentioned.
The document provides an overview of deep learning, including its past, present, and future. It discusses the concepts of artificial general intelligence, artificial superintelligence, and predictions about their development from experts like Hawking, Musk, and Gates. Key deep learning topics are summarized, such as neural networks, machine learning approaches, important algorithms and researchers, and how deep learning works.
Introduction to Recurrent Neural NetworkKnoldus Inc.
The document provides an introduction to recurrent neural networks (RNNs). It discusses how RNNs differ from feedforward neural networks in that they have internal memory and can use their output from the previous time step as input. This allows RNNs to process sequential data like time series. The document outlines some common RNN types and explains the vanishing gradient problem that can occur in RNNs due to multiplication of small gradient values over many time steps. It discusses solutions to this problem like LSTMs and techniques like weight initialization and gradient clipping.
Introduction For seq2seq(sequence to sequence) and RNNHye-min Ahn
This is my slides for introducing sequence to sequence model and Recurrent Neural Network(RNN) to my laboratory colleagues.
Hyemin Ahn, @CPSLAB, Seoul National University (SNU)
Deep Learning Tutorial | Deep Learning TensorFlow | Deep Learning With Neural...Simplilearn
This Deep Learning presentation will help you in understanding what is Deep Learning, why do we need Deep learning, what is neural network, applications of Deep Learning, what is perceptron, implementing logic gates using perceptron, types of neural networks. At the end of the video, you will get introduced to TensorFlow along with a usecase implementation on recognizing hand-written digits. Deep Learning is inspired by the integral function of the human brain specific to artificial neural networks. These networks, which represent the decision-making process of the brain, use complex algorithms that process data in a non-linear way, learning in an unsupervised manner to make choices based on the input. Deep Learning, on the other hand, uses advanced computing power and special type of neural networks and applies them to large amounts of data to learn, understand, and identify complicated patterns. W will also understand neural networks and how they work in this Deep Learning tutorial video. This Deep Learning tutorial is ideal for professionals with beginner to intermediate level of experience. Now, let us dive deep into this topic and understand what Deep Learning actually is.
Below topics are explained in this Deep Learning presentation:
1. What is Deep Learning?
2. Why do we need Deep Learning?
3. What is Neural network?
4. What is Perceptron?
5. Implementing logic gates using Perceptron
6. Types of Neural networks
7. Applications of Deep Learning
8. Working of Neural network
9. Introduction to TensorFlow
10. Use case implementation using TensorFlow
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
Part 2 of the Deep Learning Fundamentals Series, this session discusses Tuning Training (including hyperparameters, overfitting/underfitting), Training Algorithms (including different learning rates, backpropagation), Optimization (including stochastic gradient descent, momentum, Nesterov Accelerated Gradient, RMSprop, Adaptive algorithms - Adam, Adadelta, etc.), and a primer on Convolutional Neural Networks. The demos included in these slides are running on Keras with TensorFlow backend on Databricks.
Our fall 12-Week Data Science bootcamp starts on Sept 21st,2015. Apply now to get a spot!
If you are hiring Data Scientists, call us at (1)888-752-7585 or reach [email protected] to share your openings and set up interviews with our excellent students.
---------------------------------------------------------------
Come join our meet-up and learn how easily you can use R for advanced Machine learning. In this meet-up, we will demonstrate how to understand and use Xgboost for Kaggle competition. Tong is in Canada and will do remote session with us through google hangout.
---------------------------------------------------------------
Speaker Bio:
Tong is a data scientist in Supstat Inc and also a master students of Data Mining. He has been an active R programmer and developer for 5 years. He is the author of the R package of XGBoost, one of the most popular and contest-winning tools on kaggle.com nowadays.
Pre-requisite(if any): R /Calculus
Preparation: A laptop with R installed. Windows users might need to have RTools installed as well.
Agenda:
Introduction of Xgboost
Real World Application
Model Specification
Parameter Introduction
Advanced Features
Kaggle Winning Solution
Event arrangement:
6:45pm Doors open. Come early to network, grab a beer and settle in.
7:00-9:00pm XgBoost Demo
Reference:
https://github.com/dmlc/xgboost
This document provides an agenda for a presentation on deep learning, neural networks, convolutional neural networks, and interesting applications. The presentation will include introductions to deep learning and how it differs from traditional machine learning by learning feature representations from data. It will cover the history of neural networks and breakthroughs that enabled training of deeper models. Convolutional neural network architectures will be overviewed, including convolutional, pooling, and dense layers. Applications like recommendation systems, natural language processing, and computer vision will also be discussed. There will be a question and answer section.
Part 1 of the Deep Learning Fundamentals Series, this session discusses the use cases and scenarios surrounding Deep Learning and AI; reviews the fundamentals of artificial neural networks (ANNs) and perceptrons; discuss the basics around optimization beginning with the cost function, gradient descent, and backpropagation; and activation functions (including Sigmoid, TanH, and ReLU). The demos included in these slides are running on Keras with TensorFlow backend on Databricks.
Basic concept of Deep Learning with explaining its structure and backpropagation method and understanding autograd in PyTorch. (+ Data parallism in PyTorch)
Recurrent Neural Networks have shown to be very powerful models as they can propagate context over several time steps. Due to this they can be applied effectively for addressing several problems in Natural Language Processing, such as Language Modelling, Tagging problems, Speech Recognition etc. In this presentation we introduce the basic RNN model and discuss the vanishing gradient problem. We describe LSTM (Long Short Term Memory) and Gated Recurrent Units (GRU). We also discuss Bidirectional RNN with an example. RNN architectures can be considered as deep learning systems where the number of time steps can be considered as the depth of the network. It is also possible to build the RNN with multiple hidden layers, each having recurrent connections from the previous time steps that represent the abstraction both in time and space.
- Naive Bayes is a classification technique based on Bayes' theorem that uses "naive" independence assumptions. It is easy to build and can perform well even with large datasets.
- It works by calculating the posterior probability for each class given predictor values using the Bayes theorem and independence assumptions between predictors. The class with the highest posterior probability is predicted.
- It is commonly used for text classification, spam filtering, and sentiment analysis due to its fast performance and high success rates compared to other algorithms.
The document discusses Long Short Term Memory (LSTM) networks, which are a type of recurrent neural network capable of learning long-term dependencies. It explains that unlike standard RNNs, LSTMs use forget, input, and output gates to control the flow of information into and out of the cell state, allowing them to better capture long-range temporal dependencies in sequential data like text, audio, and time-series data. The document provides details on how LSTM gates work and how LSTMs can be used for applications involving sequential data like machine translation and question answering.
1) Deep learning is a type of machine learning that uses neural networks with many layers to learn representations of data with multiple levels of abstraction.
2) Deep learning techniques include unsupervised pretrained networks, convolutional neural networks, recurrent neural networks, and recursive neural networks.
3) The advantages of deep learning include automatic feature extraction from raw data with minimal human effort, and surpassing conventional machine learning algorithms in accuracy across many data types.
Tijmen Blankenvoort, co-founder Scyfer BV, presentation at Artificial Intelligence Meetup 15-1-2014. Introduction into Neural Networks and Deep Learning.
Artificial Intelligence, Machine Learning, Deep Learning
The 5 myths of AI
Deep Learning in action
Basics of Deep Learning
NVIDIA Volta V100 and AWS P3
Machine learning helps predict behavior and recognize patterns that humans cannot by learning from data without relying on programmed rules. It is an algorithmic approach that differs from statistical modeling which formalizes relationships through mathematical equations. Machine learning is a part of the broader field of artificial intelligence which aims to develop systems that can act and respond intelligently like humans. The machine learning workflow involves collecting and preprocessing data, selecting algorithms, training models, and evaluating performance. Common machine learning algorithms include supervised learning, unsupervised learning, reinforcement learning, and deep learning. Popular tools for machine learning include Python, R, TensorFlow, and Spark.
The document provides an overview of convolutional neural networks (CNNs) and their layers. It begins with an introduction to CNNs, noting they are a type of neural network designed to process 2D inputs like images. It then discusses the typical CNN architecture of convolutional layers followed by pooling and fully connected layers. The document explains how CNNs work using a simple example of classifying handwritten X and O characters. It provides details on the different layer types, including convolutional layers which identify patterns using small filters, and pooling layers which downsample the inputs.
Deep learning is introduced along with its applications and key players in the field. The document discusses the problem space of inputs and outputs for deep learning systems. It describes what deep learning is, providing definitions and explaining the rise of neural networks. Key deep learning architectures like convolutional neural networks are overviewed along with a brief history and motivations for deep learning.
Deep Learning - The Past, Present and Future of Artificial IntelligenceLukas Masuch
The document provides an overview of deep learning, including its history, key concepts, applications, and recent advances. It discusses the evolution of deep learning techniques like convolutional neural networks, recurrent neural networks, generative adversarial networks, and their applications in computer vision, natural language processing, and games. Examples include deep learning for image recognition, generation, segmentation, captioning, and more.
Automatic Attendace using convolutional neural network Face Recognitionvatsal199567
Automatic Attendance System will recognize the face of the student through the camera in the class and mark the attendance. It was built in Python with Machine Learning.
MachinaFiesta: A Vision into Machine Learning 🚀GDSCNiT
🕵️♂️ Embark on an exhilarating journey into the realm of Machine learning and Generative AI with MachinaFiesta! 🚀. Join us for MachinaFiesta, a two-hour event exploring the fascinating world of machine learning and generative AI where you can Vision, Innovate and learn new technologies.
Slide contets:
🎤 Brief introduction to the agenda and speakers of the event
🌐 Get to know the importance and future prospects of machine learning
🧠 Interactive session on core machine learning concepts
🚀 Exploration of cutting-edge generative AI advancements
🤖 Introduction to Gemini, the open-source factual language model
🤔Discussion on Gemini's capabilities and potential applications in research and development
The Backbone of Modern AI Models" The architecture of TransformersJoshiniM2
The Backbone of Modern AI Models"
The architecture of Transformers
How they differ from traditional neural networks
Their role in Foundation Models
Applications in NLP (e.g., BERT, GPT)
Why they are essential for Large Language Models (LLMs)
Traditional Machine Learning had used handwritten features and modality-specific machine learning to classify images, text or recognize voices. Deep learning / Neural network identifies features and finds different patterns automatically. Time to build these complex tasks has been drastically reduced and accuracy has exponentially increased because of advancements in Deep learning. Neural networks have been partly inspired from how 86 billion neurons work in a human and become more of a mathematical and a computer problem. We will see by the end of the blog how neural networks can be intuitively understood and implemented as a set of matrix multiplications, cost function, and optimization algorithms.
Deep neural networks learn hierarchical representations of data through multiple layers of feature extraction. Lower layers identify low-level features like edges while higher layers integrate these into more complex patterns and objects. Deep learning models are trained on large labeled datasets by presenting examples, calculating errors, and adjusting weights to minimize errors over many iterations. Deep learning has achieved human-level performance on tasks like image recognition due to its ability to leverage large amounts of training data and learn representations automatically rather than relying on manually designed features.
An Introduction to Deep Learning (May 2018)Julien SIMON
This document provides an introduction to deep learning, including common network architectures and use cases. It defines artificial intelligence, machine learning, and deep learning. It discusses how neural networks are trained using stochastic gradient descent and backpropagation to minimize loss and optimize weights. Common network types are described, such as convolutional neural networks for image recognition and LSTM networks for sequence prediction. Examples of deep learning applications include machine translation, object detection, segmentation, and generation of images, text, and video. Resources for learning more about deep learning are provided.
This document provides an overview of machine learning and deep learning concepts. It begins with an introduction to machine learning basics, including supervised and unsupervised learning. It then discusses deep learning, why it is useful, and its main components like activation functions, optimizers, and regularization methods. The document explains deep neural network architecture including convolutional neural networks. It provides examples of convolutional and max pooling layers and how they help reduce parameters in neural networks.
The document discusses machine learning and learning agents in three main points:
1. It defines machine learning and discusses different types of machine learning tasks like supervised, unsupervised, and reinforcement learning.
2. It explains the key differences between traditional machine learning approaches and learning agents, noting that learning is one of many goals for agents and must be integrated with other agent functions.
3. It discusses different challenges of integrating machine learning into intelligent agents, such as balancing learning with recall of existing knowledge and addressing time constraints on learning from the environment.
Deep learning techniques like convolutional neural networks (CNNs) and deep neural networks have achieved human-level performance on certain tasks. Pioneers in the field include Geoffrey Hinton, who co-invented backpropagation, Yann LeCun who developed CNNs for image recognition, and Andrew Ng who helped apply these techniques at companies like Baidu and Coursera. Deep learning is now widely used for applications such as image recognition, speech recognition, and distinguishing objects like dogs from cats, often outperforming previous machine learning methods.
Deep Learning: concepts and use cases (October 2018)Julien SIMON
An introduction to Deep Learning theory
Neurons & Neural Networks
The Training Process
Backpropagation
Optimizers
Common network architectures and use cases
Convolutional Neural Networks
Recurrent Neural Networks
Long Short Term Memory Networks
Generative Adversarial Networks
Getting started
Training machine learning deep learning 2017Iwan Sofana
This document discusses deep learning and neural networks. It begins with a brief history of neural networks, from the earliest Perceptron algorithm in 1958 to modern developments enabled by increased computational power and data. Deep learning uses neural networks with multiple hidden layers to automatically learn representations of data and hierarchical feature detectors. Examples are given of applying deep learning to tasks like image recognition. The document outlines challenges of deep learning like the large amount of training required and complexity of modeling real-world behaviors.
Understanding Deep Learning & Parameter Tuning with MXnet, H2o Package in RManish Saraswat
Simple guide which explains deep learning and neural network with hands on experience in R using MXnet and H2o package. It also explains gradient descent and backpropagation algorithm.
Complete tutorial: http://blog.hackerearth.com/understanding-deep-learning-parameter-tuning-with-mxnet-h2o-package-r
introduction to embedded system presentationAmr Rashed
An embedded system is a type of electronic system programmed to perform specific tasks. It contains hardware and software components that work together to perform functions like displaying time on a watch or washing clothes in a washing machine. Key components of an embedded system include a processor, memory, input/output interfaces and application software. Embedded systems have become more advanced over time, evolving from using vacuum tubes and transistors to today's microcontrollers and microprocessors. They provide advantages like small size, low power consumption and low cost. Common applications include consumer electronics, automobiles, industrial automation and medical devices.
This document summarizes key concepts from Chapter 5 on counting principles, permutations, and combinations. It introduces the product rule and sum rule for counting the number of possible outcomes of multi-step processes. It then covers permutations, which are ordered arrangements, and combinations, which are unordered selections of elements from a set. Examples are provided to illustrate calculating permutations and combinations using formulas like P(n,r) and C(n,r). The chapter also discusses proof techniques like direct proof, proof by contradiction, and proof by contraposition.
This document is the chapter outline for Chapter 8 of a course on Discrete Structures. The chapter covers properties of relations, combining relations, matrix representations of relations, representing relations using digraphs, and equivalence relations. It includes 9 reference YouTube videos providing additional information on these topics related to relations and discrete mathematics.
Discrete Math Chapter 1 :The Foundations: Logic and ProofsAmr Rashed
The document describes Chapter 1 of a textbook on discrete mathematics and its applications. Chapter 1 covers propositional logic, propositional equivalences, predicates and quantifiers, and nested quantifiers. It defines basic concepts such as propositional variables, logical operators, truth tables, logical equivalence, predicates, quantifiers, and translating between logical expressions and English sentences. Examples are provided to illustrate different logical equivalences and how to negate quantified statements. The chapter introduces key foundations of logic and proofs that are important for discrete mathematics.
1) The document summarizes key concepts from a discrete mathematics course including sets, operations on sets, functions, sequences, and sums.
2) It covers topics like basic set theory, operations on sets like union and intersection, subsets, power sets, Cartesian products, and cardinality.
3) Examples are provided to illustrate concepts like Venn diagrams, disjoint sets, complements, and set differences.
The document summarizes key concepts in discrete mathematics including sets, operations on sets, functions, sequences, and counting techniques. It defines what a set is, ways to describe sets, and set operations like unions and intersections. Examples are given of common sets like integers, rational numbers, and real numbers. Subsets, the empty set, cardinality (size) of sets, and Venn diagrams are also explained.
Implementation of DNA sequence alignment algorithms using Fpga ,ML,and CNNAmr Rashed
The document discusses implementing DNA/RNA sequence alignment algorithms using FPGA. It proposes a hardware implementation that relies on complete parallelization of algorithms like Smith-Waterman and Needleman-Wunsch under certain limitations. A lookup table is used to accelerate the algorithms in O(N/4) time. Fifty-four Boolean functions are derived for parallel implementation. The implementation represents DNA/RNA sequences as combinations of 4 nucleotides and is applicable to both software and hardware.
This document provides an overview of information security topics including:
- Types of security attacks such as those from internal and external attackers.
- Key security concepts like confidentiality, integrity, and availability.
- Examples of security violations involving unauthorized access or modification of files.
- The importance of considering security attacks, mechanisms, and services as major axes in network security.
Machine learning workshop using Orange datamining frameworkAmr Rashed
Machine learning workshop using Orange
youtube video
https://youtu.be/wpgQY5f2hOo
Topic: Data mining, analysis, and visualization Using Python-Orange
Start Time : Mar 27, 2021 08:30 PM
Meeting Recording:
https://zoom.us/rec/share/esp-FwuaZs3ekc-yYNK74EV7Jn-TSM1TpmT2fTbe8Oy99MKmsdDhQigRneEyQaM-.JNssJnqQqtrAVgQO
This document provides an overview and introduction to deep learning. It discusses motivations for deep learning such as its powerful learning capabilities. It then covers deep learning basics like neural networks, neurons, training processes, and gradient descent. It also discusses different network architectures like convolutional neural networks and recurrent neural networks. Finally, it describes various deep learning applications, tools, and key researchers and companies in the field.
The document contains 20 MATLAB programs demonstrating various plotting and data visualization techniques including:
- Plotting sinusoidal waves and applying half/full wave rectification
- Converting between polar and Cartesian coordinates and plotting circles
- Generating and plotting cylinder surfaces
- Using EZ functions like ezplot, ezsurf, and ezpolar to plot functions
- Taking derivatives and plotting functions and their derivatives
- Plotting noise signals and calculating rate of change over time
- Generating 2D and 3D surfaces from gridded data
- Plotting polar plots and converting to Cartesian coordinates
- Applying FFT and filtering to decompose a signal into frequency components
- Using pie charts and 3D pie plots to
This document presents a fast algorithm for license plate detection. It begins with an introduction that outlines the need for automatic license plate recognition systems. It then discusses previous work in the area and the challenges involved. The proposed technique is divided into four main parts: histogram equalization, removal of border and background, image segmentation, and license plate detection using feature extraction, principal component analysis, and artificial neural networks. Test results on a dataset of 30 images achieved a 93.33% detection rate. Future work involves implementing the neural network classifier on an FPGA for increased speed.
This document provides an introduction and overview of various digital logic and programmable devices including VHDL, microcontrollers, DSPs, PLCs, PLDs, ASICs, and FPGAs. It defines these terms and describes the basic architecture and applications of each technology. References and resources for further reading are also provided.
Analysis of reinforced concrete deep beam is based on simplified approximate method due to the complexity of the exact analysis. The complexity is due to a number of parameters affecting its response. To evaluate some of this parameters, finite element study of the structural behavior of the reinforced self-compacting concrete deep beam was carried out using Abaqus finite element modeling tool. The model was validated against experimental data from the literature. The parametric effects of varied concrete compressive strength, vertical web reinforcement ratio and horizontal web reinforcement ratio on the beam were tested on eight (8) different specimens under four points loads. The results of the validation work showed good agreement with the experimental studies. The parametric study revealed that the concrete compressive strength most significantly influenced the specimens’ response with the average of 41.1% and 49 % increment in the diagonal cracking and ultimate load respectively due to doubling of concrete compressive strength. Although the increase in horizontal web reinforcement ratio from 0.31 % to 0.63 % lead to average of 6.24 % increment on the diagonal cracking load, it does not influence the ultimate strength and the load-deflection response of the beams. Similar variation in vertical web reinforcement ratio leads to an average of 2.4 % and 15 % increment in cracking and ultimate load respectively with no appreciable effect on the load-deflection response.
Fluid mechanics is the branch of physics concerned with the mechanics of fluids (liquids, gases, and plasmas) and the forces on them. Originally applied to water (hydromechanics), it found applications in a wide range of disciplines, including mechanical, aerospace, civil, chemical, and biomedical engineering, as well as geophysics, oceanography, meteorology, astrophysics, and biology.
It can be divided into fluid statics, the study of various fluids at rest, and fluid dynamics.
Fluid statics, also known as hydrostatics, is the study of fluids at rest, specifically when there's no relative motion between fluid particles. It focuses on the conditions under which fluids are in stable equilibrium and doesn't involve fluid motion.
Fluid kinematics is the branch of fluid mechanics that focuses on describing and analyzing the motion of fluids, such as liquids and gases, without considering the forces that cause the motion. It deals with the geometrical and temporal aspects of fluid flow, including velocity and acceleration. Fluid dynamics, on the other hand, considers the forces acting on the fluid.
Fluid dynamics is the study of the effect of forces on fluid motion. It is a branch of continuum mechanics, a subject which models matter without using the information that it is made out of atoms; that is, it models matter from a macroscopic viewpoint rather than from microscopic.
Fluid mechanics, especially fluid dynamics, is an active field of research, typically mathematically complex. Many problems are partly or wholly unsolved and are best addressed by numerical methods, typically using computers. A modern discipline, called computational fluid dynamics (CFD), is devoted to this approach. Particle image velocimetry, an experimental method for visualizing and analyzing fluid flow, also takes advantage of the highly visual nature of fluid flow.
Fundamentally, every fluid mechanical system is assumed to obey the basic laws :
Conservation of mass
Conservation of energy
Conservation of momentum
The continuum assumption
For example, the assumption that mass is conserved means that for any fixed control volume (for example, a spherical volume)—enclosed by a control surface—the rate of change of the mass contained in that volume is equal to the rate at which mass is passing through the surface from outside to inside, minus the rate at which mass is passing from inside to outside. This can be expressed as an equation in integral form over the control volume.
The continuum assumption is an idealization of continuum mechanics under which fluids can be treated as continuous, even though, on a microscopic scale, they are composed of molecules. Under the continuum assumption, macroscopic (observed/measurable) properties such as density, pressure, temperature, and bulk velocity are taken to be well-defined at "infinitesimal" volume elements—small in comparison to the characteristic length scale of the system, but large in comparison to molecular length scale
Sorting Order and Stability in Sorting.
Concept of Internal and External Sorting.
Bubble Sort,
Insertion Sort,
Selection Sort,
Quick Sort and
Merge Sort,
Radix Sort, and
Shell Sort,
External Sorting, Time complexity analysis of Sorting Algorithms.
Value Stream Mapping Worskshops for Intelligent Continuous SecurityMarc Hornbeek
This presentation provides detailed guidance and tools for conducting Current State and Future State Value Stream Mapping workshops for Intelligent Continuous Security.
The Fluke 925 is a vane anemometer, a handheld device designed to measure wind speed, air flow (volume), and temperature. It features a separate sensor and display unit, allowing greater flexibility and ease of use in tight or hard-to-reach spaces. The Fluke 925 is particularly suitable for HVAC (heating, ventilation, and air conditioning) maintenance in both residential and commercial buildings, offering a durable and cost-effective solution for routine airflow diagnostics.
ELectronics Boards & Product Testing_Shiju.pdfShiju Jacob
This presentation provides a high level insight about DFT analysis and test coverage calculation, finalizing test strategy, and types of tests at different levels of the product.
"Feed Water Heaters in Thermal Power Plants: Types, Working, and Efficiency G...Infopitaara
A feed water heater is a device used in power plants to preheat water before it enters the boiler. It plays a critical role in improving the overall efficiency of the power generation process, especially in thermal power plants.
🔧 Function of a Feed Water Heater:
It uses steam extracted from the turbine to preheat the feed water.
This reduces the fuel required to convert water into steam in the boiler.
It supports Regenerative Rankine Cycle, increasing plant efficiency.
🔍 Types of Feed Water Heaters:
Open Feed Water Heater (Direct Contact)
Steam and water come into direct contact.
Mixing occurs, and heat is transferred directly.
Common in low-pressure stages.
Closed Feed Water Heater (Surface Type)
Steam and water are separated by tubes.
Heat is transferred through tube walls.
Common in high-pressure systems.
⚙️ Advantages:
Improves thermal efficiency.
Reduces fuel consumption.
Lowers thermal stress on boiler components.
Minimizes corrosion by removing dissolved gases.
ADVXAI IN MALWARE ANALYSIS FRAMEWORK: BALANCING EXPLAINABILITY WITH SECURITYijscai
With the increased use of Artificial Intelligence (AI) in malware analysis there is also an increased need to
understand the decisions models make when identifying malicious artifacts. Explainable AI (XAI) becomes
the answer to interpreting the decision-making process that AI malware analysis models use to determine
malicious benign samples to gain trust that in a production environment, the system is able to catch
malware. With any cyber innovation brings a new set of challenges and literature soon came out about XAI
as a new attack vector. Adversarial XAI (AdvXAI) is a relatively new concept but with AI applications in
many sectors, it is crucial to quickly respond to the attack surface that it creates. This paper seeks to
conceptualize a theoretical framework focused on addressing AdvXAI in malware analysis in an effort to
balance explainability with security. Following this framework, designing a machine with an AI malware
detection and analysis model will ensure that it can effectively analyze malware, explain how it came to its
decision, and be built securely to avoid adversarial attacks and manipulations. The framework focuses on
choosing malware datasets to train the model, choosing the AI model, choosing an XAI technique,
implementing AdvXAI defensive measures, and continually evaluating the model. This framework will
significantly contribute to automated malware detection and XAI efforts allowing for secure systems that
are resilient to adversarial attacks.
☁️ GDG Cloud Munich: Build With AI Workshop - Introduction to Vertex AI! ☁️
Join us for an exciting #BuildWithAi workshop on the 28th of April, 2025 at the Google Office in Munich!
Dive into the world of AI with our "Introduction to Vertex AI" session, presented by Google Cloud expert Randy Gupta.
In tube drawing process, a tube is pulled out through a die and a plug to reduce its diameter and thickness as per the requirement. Dimensional accuracy of cold drawn tubes plays a vital role in the further quality of end products and controlling rejection in manufacturing processes of these end products. Springback phenomenon is the elastic strain recovery after removal of forming loads, causes geometrical inaccuracies in drawn tubes. Further, this leads to difficulty in achieving close dimensional tolerances. In the present work springback of EN 8 D tube material is studied for various cold drawing parameters. The process parameters in this work include die semi-angle, land width and drawing speed. The experimentation is done using Taguchi’s L36 orthogonal array, and then optimization is done in data analysis software Minitab 17. The results of ANOVA shows that 15 degrees die semi-angle,5 mm land width and 6 m/min drawing speed yields least springback. Furthermore, optimization algorithms named Particle Swarm Optimization (PSO), Simulated Annealing (SA) and Genetic Algorithm (GA) are applied which shows that 15 degrees die semi-angle, 10 mm land width and 8 m/min drawing speed results in minimal springback with almost 10.5 % improvement. Finally, the results of experimentation are validated with Finite Element Analysis technique using ANSYS.
"Boiler Feed Pump (BFP): Working, Applications, Advantages, and Limitations E...Infopitaara
A Boiler Feed Pump (BFP) is a critical component in thermal power plants. It supplies high-pressure water (feedwater) to the boiler, ensuring continuous steam generation.
⚙️ How a Boiler Feed Pump Works
Water Collection:
Feedwater is collected from the deaerator or feedwater tank.
Pressurization:
The pump increases water pressure using multiple impellers/stages in centrifugal types.
Discharge to Boiler:
Pressurized water is then supplied to the boiler drum or economizer section, depending on design.
🌀 Types of Boiler Feed Pumps
Centrifugal Pumps (most common):
Multistage for higher pressure.
Used in large thermal power stations.
Positive Displacement Pumps (less common):
For smaller or specific applications.
Precise flow control but less efficient for large volumes.
🛠️ Key Operations and Controls
Recirculation Line: Protects the pump from overheating at low flow.
Throttle Valve: Regulates flow based on boiler demand.
Control System: Often automated via DCS/PLC for variable load conditions.
Sealing & Cooling Systems: Prevent leakage and maintain pump health.
⚠️ Common BFP Issues
Cavitation due to low NPSH (Net Positive Suction Head).
Seal or bearing failure.
Overheating from improper flow or recirculation.
2. Deep Learning
DEEP LEANING in Bioinformatics, Conclusion
RECURRENT NN,DEEP LEARNING TOOLS
Types of Networks , Convolution Neural Networks
DEEP NN ARCHITECTURE, PROBLEM SPACE
WHAT IS DEEP LEARNING, DEEP LEARNING BASICS
BIG PLAYERS, APPLICATIONS
A Brief History, MACHINE LEARNING BASICS
MOTIVATIONS, WHY DEEP NN
AGENDA
3. Motivations
I have worked all my life in machine learning, and I’ve
never seen one algorithm knock over benchmarks like
deep learning . Andrew Ng(Stanford & Baidu)
Deep learning is an algorithm which has no theoretical
limitations of what it can learn ;the more data you give
and the more computational time you provide ;the
better it is-Geoffrey Hilton(Google)
Human-level artificial intelligence has the potential to
help humanity thrive more than any invention that has
come before it –Dileep George(Co-Founder Vicarious)
For a very long time it will be a complementary tool that
human scientists and human experts can use to help
them with the things that humans are not naturally
good- Demis Hassabis (Co-Founder Deep Mind)
6. The Problem Space
0 Image classification is the task of taking an input image and
outputting a class (a cat, dog, etc) or a probability of classes that
best describes the image.
0 For humans, this task of recognition is one of the first skills we
learn from the moment we are born .
0 Without even thinking twice, we’re able to quickly and seamlessly
identify the environment and objects that surround us.
0 When we see an image we are able to immediately characterize the
scene and give each object a label, all without even consciously
noticing.
0 These skills of being able to quickly recognize patterns, generalize
from prior knowledge, and adapt to different image environments
are ones that we do not share with our fellow machines.
7. The Problem Space : inputs & outputs
0 Depending on the resolution and size of the image, it will see a 32 x
32 x 3 array of numbers
0 These numbers, while meaningless to us when we perform image
classification, are the only inputs available to the computer.
0 The idea is that you give the computer this array of numbers and it
will output numbers that describe the probability of the image
being a certain class (.80 for cat, .15 for dog, .05 for bird, etc).
8. The Problem Space :What We Want the Computer to Do
0 To be able to differentiate between all the images it’s given
and figure out the unique features that make a dog a dog or
that make a cat a cat.
0 This is the process that goes on in our minds
subconsciously as well.
0 When we look at a picture of a dog, we can classify it as
such if the picture has identifiable features such as paws or
4 legs.
0 In a similar way, the computer is able perform image
classification by looking for low level features such as
edges and curves, and then building up to more abstract
concepts through a series of convolutional layers.
9. Why Deep Neural Network
- Imagine you have extracted features from sensors .
- The dimension of each sample is around 800
- You have 70,000 samples (trial)
- What method would you apply?
- You may have several ways to do it
0 Reduce the dimension from 800 to 40 by using a feature selection or
dim. reduction technique
0 What you did here is “Finding a good representation”
0 - Then, you may apply a classification methods to classify 10 classes
But, what if
- You have no idea for feature selection?
- The dimension is much higher than 800 and you have more classes.
10. Why Deep Neural Network
0 Drawbacks -Back-Propagation:
0 Scalability -does not scale well over multiple layers.
0 very slow to converge.
0“Vanishing gradient problem “ errors shrink exponentially with the
number of layers
0Thus makes poor use of many layers.
0This is the reason most feed forward NN have only three layers.
0 It got stuck at local optima
0 When the network outputs have not got their desired signals, the
activation functions in the hidden layer are saturating.
0These were often surprisingly good but there was no good theory.
0 Traditional Machine Learning doesn’t work well when
features can’t be explicitly defined.
11. A Brief History
0 2012 was the first year that neural nets grew to prominence
as Alex Krizhevsky used them to win that year’s ImageNet
competition (basically, the annual Olympics of computer
vision), dropping the classification error record from 26% to
15%, an astounding improvement at the time.
12. Machine Learning –Basics
Introduction
Machine Learning is a type of Artificial
Intelligence that provides computers with the
ability to learn without being explicitly
programmed. Provides various techniques that
can learn from and make predictions on data.
13. Supervised Learning : Learning with a labeled
training set
Example : e-mail spam detector with training set of
already labeled emails
Unsupervised Learning :Discovering patterns in
unlabeled data
Example :cluster similar documents based on the text
content
Reinforcement learning : learning based on feedback
or reward.
Example :learn to play chess by wining or losing
Machine Learning –Basics
Learning Approach
15. Big Players :Superstar Researcher
Jeoferry Hilton : University of Toronto
&Google
Yann Lecun : New Yourk University
&Facebook
Andrew Ng : Stanford & Baidu
Yoshua Bengio : University of Montreal
Jurgen schmidhuber : Swiss AI Lab &
NNAISENSE
20. What is Deep Learning(DL)
Part or a powerful class of the machine learning field of
learning representations of data. Exceptional effective at
learning patterns.
Utilize learning algorithms that derive meaning out of data
by using hierarchy of multiple layers that mimic the neural
networks of our brain.
If you provide the system tons of information , it begins to
understand it and respond in useful ways.
Stacked “Neural Network”. Is usually indicates “Deep Neural Network”.
21. What is Deep Learning (DL)
• Collection of simple, trainable mathematical functions.
• Learning methods which have deep architecture.
• It often allows end-to-end learning.
• It automatically finds intermediate representation. Thus, it can
be regarded as a representation learning.
22. Deep Learning Basics :Neuron
An artificial neuron contains a nonlinear activation function and has several
incoming and outgoing weighted connections.
Neurons are trained filters and detect specific features or patterns (e.g.
edge, nose) by receiving weighted input, transforming it with the activation
function and passing it to the outgoing connections
23. Commonalities with real brains:
● Each neuron is connected to a small subset of other neurons.
● Based on what it sees, it decides what it wants to say.
● Neurons learn to cooperate to accomplish the task.
• The vental pathway in the visual cortex has multiple stages
• There exist a lot of intermediate representations
Deep Learning Basics :Neuron
24. Deep Learning Basics :Training process
Learns by generating an error signal that measures the difference between the
predictions of the network and the desired values and then using this error signal
to change the weights (or parameters) so the predictions get more accurate.
25. Deep Learning Basics :Gradient Descent
Gradient Descent/optimization finds the (local)the minimum of the cost function(used
to calculate the output error) and is used to adjust the weights.
Curve represents is the network's error relative to the position of a single weight.
X axis weight Y axiserror
Oversimplified Gradient Descent:
Calculate slope at current position
• If slope is negative, move right
• If slope is positive, move left
• (Repeat until slope == 0)
26. Deep Learning Basics :Gradient Descent Problems
Problem 1: When slopes are too big Solution 1: Make Slopes Smaller
Problem 2: Local Minimums-Solution 2: Multiple Random Starting States
27. Deep Neural Network :Architecture
A Deep Neural Network consists of a hierarchy of layers, whereby each layer
transforms the input data into more abstract representations (e.g edge -
>nose -> face). The output layer combines those features to make predictions.
28. Deep Neural Network :Architecture
Consists of one input ,one output and multiple fully-connected hidden layers in
between. Each layer is represented as a series of neurons and progressively extracts
higher and higher-level features of the input until the final layer essentially makes a
decision about what the input shows. The more layer the network has, the higher-level
features it will learn.
30. Types of Networks used for Deep Learning
Convolutional Neural Networks(Convnet,CNN)
Recurrent Neural Networks(RNN)
Long Short Term Memory (LSTM) networks
Deep/Restricted Boltzmann Machines (RBM)
Deep Q-networks
Deep Belief Networks(DBN)
Deep Stacking Networks
31. Convolution Neural Networks(CNN):Basic Idea
0 This idea was expanded upon by a fascinating experiment by
Hubel and Wiesel in 1962 where they showed that some
individual neuronal cells in the brain responded (or fired)
only in the presence of edges of a certain orientation.
0 For example, some neurons fired when exposed to vertical
edges and some when shown horizontal or diagonal edges.
Hubel and Wiesel found out that all of these neurons were
organized in a columnar architecture and that together, they
were able to produce visual perception.
0 This idea of specialized components inside of a system
having specific tasks (the neuronal cells in the visual cortex
looking for specific characteristics) is one that machines use
as well, and is the basis behind CNNs.
32. Convolution Neural Networks(CNN)
• Convolutional neural networks learn a complex representation of visual data
using vast amounts of data .they are inspired by human visual system and learn
multiple layers of transformations , which are applied on top of each other to
extract progressively more sophisticated representation of the input .
DEFENITION
• Inspired by the visual cortex and Pioneered by Yann Lecun (NYU).
• CNN have multiple types of layers ,the first of which is the Convolutional
layer.
Notes
34. CNN Structure :First Layer – Convolution
Convolution layer is a feature detector that auto magically learns to filter out
not needed information from an input by using convolution kernel.
37. Examples of Actual Visualizations
End-to-End Learning
A pipeline of successive Layers
• Layers produce features of higher and higher abstractions
– Initial Layer capture low-level features (e.g. edges or corners)
– Middle Layer capture mid-level features (object parts e.g. circles, squares, textures)
– Last Layer capture high level, class specific features (object model e.g. face detector)
• Preferably, input as raw as possible
– Pixels for computer vision, words for NLP.
38. Batch Normalization Layer
0 Batch Normalization is a technique to provide any layer in a Neural Network with
inputs that are zero mean/unit variance .
0 Use batch normalization layers between convolutional layers and nonlinearities such as
ReLU layers to speed up network training and reduce the sensitivity to network
initialization.
0 The layer first normalizes the activations of each channel by subtracting the mini-batch
mean and dividing by the mini-batch standard deviation. Then, the layer shifts the input
by an offset β and scales it by a scale factor γ. β and γ are themselves learnable
parameters that are updated during network training.
39. Nonlinear Activation Function(Relu)
•Rectified Linear Unit (ReLU) module .
•Activation function 𝑎=ℎ(𝑥)=max(0,𝑥).
Most deep networks use ReLU –max(0,x)-nowadays for hidden
layers ,since it trains much faster ,is more expressive than logistic
function and prevents the gradient vanishing problem.
Non-linearity is needed to learn complex (non
linear)representations of data ,otherwise the NN would be just a
linear function .
40. Vanishing Gradient Problem
0 It is a difficulty founds in training ANN with gradient-based learning
methods and backpropagation.
0 In such methods, each of the neural network's weights receives an
update proportional to the gradient of the error function with respect
to the current weight in each iteration of training.
0 Traditional activation functions such as the hyperbolic
tangent function have gradients in the range (−1, 1), and
backpropagation computes gradients by the chain rule.
0 This has the effect of multiplying n of these small numbers to
compute gradients of the "front" layers in an n-layer network.
0 This means that the gradient (error signal) decreases exponentially
with n while the front layers train very slowly.
41. Local Response Normalization Layer
0 we want to have some kind of inhibition scheme.
0 Neurobiology concept “lateral inhibition”.
0 Useful when we are dealing with ReLU neurons(unbounded activations ).
0 Detect high frequency features with a large response.
0 If we normalize around the local neighborhood of the excited neuron, it
becomes even more sensitive as compared to its neighbors.
0 Inhibit the responses that are uniformly large in any given local
neighborhood.
0 If all the values are large, then normalizing those values will decrease all
of them.
0 Inhibition and boost the neurons with relatively larger activations.
0 Two types of normalizations available in Caffe (same and cross channel)
where K, α, and β are the hyperparameters in the
normalization, and ss is the sum of squares of the
elements in the normalization window.
42. Max and AVG Pooling Layer
0 Pooling layers compute the max or average value of a particular feature over a
region of the input data (downsizing of input images).
0 Also helps to detect objects in some unusual places and reduces memory size.
0 Aggregate multiple values into a single value
0 – Reduces the size of the layer output/input to next layer ->Faster computations
0 – Keeps most important information for the next layer
0 • Max pooling/Average pooling
43. Over Fitting
0 Over Fitting. This term refers to when a model is so tuned to the training
examples that it is not able to generalize well for the validation and test sets.
0 A symptom of over Fitting is having a model that gets 100% or 99% on the
training set, but only 50% on the test data.
0 Implement dropout layers in order to combat the problem of over fitting to the
training data.
44. Dropout Layers
0 after training, the weights of the network are so tuned to the
training examples they are given that the network doesn’t
perform well when given new examples.
0 The idea of dropout is simplistic in nature. This layer “drops out”
a random set of activations in that layer by setting them to zero.
Simple as that.
0 Benefits
0 it forces the network to be redundant. By that the network should
be able to provide the right classification or output for a specific
example even if some of the activations are dropped out.
0 It makes sure that the network isn’t getting too “fitted” to the
training data and thus helps alleviate the over fitting problem.
0 An important note is that this layer is only used during training,
and not during test time.
45. Output Layer –Soft-max Activation Function
0 For classification problems the sigmoid function can only handle two classes, which
is not what we expect.
0 The softmax function squashes the outputs of each unit to be between 0 and 1, just
like a sigmoid function.
0 But it also divides each output such that the total sum of the outputs is equal to 1 .
0 The output of the softmax function is equivalent to a categorical probability
distribution, it tells you the probability that any of the classes are true.
46. Output Layer –Regression
0 You can also use ConvNets for regression problems, where the
target (output) variable is continuous.
0 In such cases, a regression output layer must follow the final
fully connected layer. You can create a regression layer using
the regressionLayer function.
0 The default loss function for a regression layer is the mean
squared error.
0 where ti is the target output, and yi is the network’s prediction
for the response variable corresponding to observation i.
47. Transfer Learning
0 Transfer learning is the process of taking a pre-trained model (the
weights and parameters of a network that has been trained on a
large dataset by somebody else) and “fine-tuning” the model with
your own dataset.
0 Pre-trained model will act as a feature extractor.
0 Freeze the weights of all the other layers .
0 Remove the last layer of the network and replace it with your own
classifier.
0 Train the network normally.
48. Cont.:Transfer Learning
• Transfer from A (image recognition) to B (radiology images)
• When transfer learning makes sense
1. Task A and B have same input X.
2. You have a lot more data for Task A than Task B.
3. Low level features from A could be helpful for learning B.
4. Weights initialization.
• When transfer learning doesn’t makes sense
1. You have a lot more data for Task B than Task A.
49. Data Augmentation Techniques
0 Approaches that alter the training data in ways that
change the array representation while keeping the label
the same.
0 They are a way to artificially expand your dataset. Some
popular augmentations people use are gray scales ,
horizontal flips, vertical flips, random crops, color jitters,
translations, rotations, and much more .
ZCA Whitening Random Rotations Random shift Random flipExample MNIST
images
50. Recurrent Neural Network (RNN)
RNNs are general computers which can learn algorithms to map
input sequences to output sequences (flexible –sized vectors).
The output vector’s contents influenced by the entire history of
input.
Applications
• In time series prediction , adaptive robotics, handwriting
recognition ,image classification, speech recognition,
stock market prediction, and other sequence learning
problems .every thing can be processed sequentially
53. Usage Requirements
Large data set with good quality (input –output
mapping)
Measurable and describable goals (define the
cost)
Enough computing power(AWS GPU Instance)
Excels in tasks where the basic unit (pixel, word)
has very little meaning in itself, but the
combination of such units has a useful meaning.
54. Deep Learning : Benefits
Robust
• No need to design the features ahead of time –features are
automatically learned to be optimal for the task at hand
• Robustness to natural variations in the data is automatically learned
Generalizable
• The same neural network approach can be used for many different
applications and data types
Scalable
• Performance improves with more data ,method is massively
parallelizable
55. Deep Learning: Weakness
1
• Deep learning requires a large dataset, hence long training period.
2
• In term of cost, Machine learning methods like SVM and other tree
ensembles are very easily deployed even by relative machine learning
novices and can usually get you reasonably good results
3
• Deep learning methods tends to learn everything. It’s better to encode
prior knowledge about structure of images (or audio or text).
4
• The learned features are often difficult to understand. Many vision
features are also not really human-understandable (e.g,
concatenations/combinations of different features).
5
• Requires a good understanding of how to model multiple modalities
with traditional tools.
56. Pre-trained Deep Learning models
• Authors :Alex Krizhevsky, Ilya Sutskever,
and Geoffrey Hinton
• Winner 2012.
• Trained the network on ImageNet data,
which contained over 15 million
• used for classification with 1000 possible
categories
• Use 11x11 sized filters in the first layer.
• Used ReLU for the nonlinearity functions .
• Used data augmentation techniques that
consisted of image translations,
horizontal reflections, and patch
extractions.
• Implemented dropout layers.
• Trained the model using batch stochastic
gradient descent, with specific values for
momentum and weight decay.
• Trained on two GTX 580 GPUs for five to
six days
• This model achieved an 15.4% error rate.
built by Matthew Zeiler and Rob Fergus
from NYU.
Winner 2013.
Very similar architecture to AlexNet, except
for a few minor modifications.
ZF Net trained on only 1.3 million images.
ZF Net used filters of size 7x7
Used ReLUs for their activation functions,
cross-entropy loss for the error function,
and trained using batch stochastic gradient
descent.
Trained on a GTX 580 GPU for twelve days.
Developed a visualization technique named
Deconvolutional Network, which helps to
examine different feature activations and
their relation to the input space. Called
“deconvnet” because it maps features to
pixels (the opposite of what a convolutional
layer does).
This model achieved an 11.2% error rate
AlexNet (2012) ZF Net (2013)
57. Pre-trained Deep Learning models
• Built by Karen Simonyan and Andrew Zisserman
of the University of Oxford
• Not a winner .
• 19 layer CNN , used 3x3 sized filters with stride
and pad of 1 , 2x2 max pooling layers with stride 2.
• The combination of two 3x3 conv layers has an
effective receptive field of 5x5.
• Decrease in the number of parameters.
• Also, with two conv layers, we’re able to use two
ReLU layers instead of one. 3 conv layers back to
back have an effective receptive field of 7x7.
• Interesting to notice that the number of filters
doubles after each max pool layer. This reinforces
the idea of shrinking spatial dimensions, but
growing depth.
• Worked well on both image classification and
localization tasks.
• used a form of localization as regression Built
model with the Caffe toolbox.
• Used scale jittering as one data augmentation
technique during training.
• Used ReLU layers after each conv layer and trained
with batch gradient descent.
• Trained on 4 Nvidia Titan Black GPUs for two to
three weeks.
• This achieved an 7.3% error rate
• GoogLeNet is a 22 layer CNN.
• Winner 2014
• Used 9 Inception modules in the whole
architecture, with over 100 layers in
total.
• No use of fully connected layers! They
use an average pool instead, to go from a
7x7x1024 volume to a 1x1x1024
volume.
• Uses 12x fewer parameters than
AlexNet.
• During testing, multiple crops of the
same image were created, fed into the
network, and the softmax probabilities
were averaged to give us the final
solution.
• Utilized concepts from R-CNN for their
detection model.
• This model places notable consideration
on memory and power usage
• Trained on “a few high-end GPUs within
a week”.
• This achieved an 6.7% error rate
VGG Net (2014) Google Net (2015)
59. Pre-trained Deep Learning models
Microsoft ResNet (2015)
input x go through conv-relu-conv series.
152 layer, “Ultra-deep” – Yann LeCun.
Winner 2015.
After only the first 2 layers, the spatial size gets compressed from an input volume of 224x224 to a 56x56
volume.
increase of layers in plain nets result in higher training and test error (author claims).
The authors believe that “it is easier to optimize the residual mapping than to optimize the original,
unreferenced mapping
The group tried a 1202-layer network, but got a lower test accuracy, presumably due to over-fitting.
Trained on an 8 GPU machine for two to three weeks.
Achieved an 3.6 % error rate.
60. 2016 LSVRC winner(CUImage team)
0 Compared with CUImage submission in ILSVRC 2015, the new components are as follows.
(1) The models are pretrained for 1000-class object detection task using the approach in [a] but adapted to
the fast-RCNN for faster detection speed.
(2) The region proposal is obtained using the improved version of CRAFT in [b].
(3) A GBD network [c] with 269 layers is fine-tuned on 200 detection classes with the gated bidirectional
network (GBD-Net), which passes messages between features from different support regions during both
feature learning and feature extraction. The GBD-Net is found to bring ~3% mAP improvement on the
baseline 269 model and ~5% mAP improvement on the Batch normalized GoogleNet.
(4) For handling their long-tail distribution problem, the 200 classes are clustered. Different from the
original implementation in [d] that learns several models, a single model is learned, where different
clusters have both shared and distinguished feature representations.
(5) Ensemble of the models using the approaches mentioned above lead to the final result in the provided
data track.
(6) For the external data track, we propose object detection with landmarks. Comparing to the standard
bounding box centric approach, our landmark centric approach provides more structural information and
can be used to improve both the localization and classification step in object detection. Based on the
landmark annotations provided in [e], we annotate 862 landmarks from 200 categories on the training set.
Then we use them to train a CNN regressor to predict landmark position and visibility of each proposal in
testing images. In the classification step, we use the landmark pooling on top of the fully convolutional
network, where features around each landmark are mapped to be a confidence score of the corresponding
category. The landmark level classification can be naturally combined with standard bounding box level
classification to get the final detection result.
(7) Ensemble of the models using the approaches mentioned above lead to the final result in the external
data track. The fastest publicly available multi-GPU caffe code is our strong support [f].
61. 2017 LSVRC winner(BDAT team)
0 Adaptive attention[1] and deep combined convolutional
models[2,3] are used for LOC task.
0 Deep residual learning for image recognition.
0 Scale[4,5,6], context[7], sampling and deep combined
convolutional networks[2,3] are considered for DET task.
0 Object density estimation is used for score re-rank
63. Solution 1:Exhaustively search for objects.
0 Problem: Extremely slow, must process tens of thousands
of candidate objects.
64. Solution 2:
Running a scanning detector is cheaper than running a recognizer, so
do that first.
1. Exhaustively search for candidate objects with a generic detector.
2. Run recognition algorithm only on candidate objects.
Problem: What about oddly-shaped
objects? Will we need to scan with windows of many different shapes?
65. Segmentation
0 Idea: If we correctly segment the image before running object
recognition, we can use our segmentations as candidate objects.
0 Advantages: Can be efficient, makes no assumptions about
object sizes or shapes.
66. Segmentation is Hard
0 As we saw in Project 1, it’s not always clear what separates an
object.
67. Segmentation is Hard
0 As we saw in Project 1, it’s not always clear what separates
an object.
68. Selective Search
Goals:
1. Detect objects at any scale.
a. Hierarchical algorithms are good at this.
2. Consider multiple grouping criteria.
a. Detect differences in color, texture, brightness, etc.
3. Be fast.
Idea: Use bottom-up grouping of image regions to generate
a hierarchy of small to large regions.
69. Selective Search
Step 1: Generate initial sub-segmentation
Goal: Generate many regions, each of which belongs to at
most one object.
Using the method described by Felzenszwalb et al. from
week 1 works well.
70. Selective Search
Step 2: Recursively combine similar regions into
larger ones.
Greedy algorithm:
1. From set of regions, choose two that are most similar.
2. Combine them into a single, larger region.
3. Repeat until only one region remains.
This yields a hierarchy of successively larger regions, just
like we want.
71. Similarity
What do we mean by “similarity”?
Goals:
1. Use multiple grouping criteria.
2. Lead to a balanced hierarchy of small to large objects.
3. Be efficient to compute: should be able to quickly combine
measurements in two regions.
Two-pronged approach:
1. Choose a color space that captures interesting things.
a. Different color spaces have different invariants, and different
responses to changes in color.
2. Choose a similarity metric for that space that captures
everything we’re interested: color, texture, size, and shape.
72. Similarity
0 RGB (red, green, blue) is a good baseline, but changes
in illumination (shadows, light intensity) affect all three
channels.
73. Similarity
0 HSV (hue, saturation, value) encodes color information in the
hue channel, which is invariant to changes in lighting.
Additionally, saturation is insensitive to shadows, and value is
insensitive to brightness changes.
74. Similarity
0 Lab uses a lightness channel and two color channels (a and
b). It’s calibrated to be perceptually uniform. Like HSV, it’s also
somewhat invariant to changes in brightness and shadow.
78. Generative Adversarial Networks
0 According to Yann LeCun, these networks could be the next big development.
0 The idea is to simultaneously train two neural nets.
0 The first one, called the Discriminator D(Y) takes an input (e.g. an image) and
outputs a scalar that indicates whether the image Y looks “natural” or not.
0 The second network is called the generator, denoted G(Z), where Z is generally
a vector randomly sampled in a simple distribution (e.g. Gaussian). The role of
the generator is to produce images so as to train the D(Y) function to take the
right shape (low values for real images, higher values for everything else).
79. Generative Adversarial Networks : Importance
0 The discriminator now is aware of the “internal
representation of the data” because it has been trained to
understand the differences between real images from the
dataset and artificially created ones.
0 It can be used as a feature extractor for CNN.
0 You can just create really cool artificial images that look
pretty natural to me.
80. Generating Image Descriptions
0 What happens when you combine CNNs with RNNs. you do get one
really amazing application.
0 Combination of CNNs and bidirectional RNNs to generate natural
language descriptions of different image regions
81. Spatial Transformer Network
0 The basic idea is that this module transforms the input image in a way so
that the subsequent layers have an easier time making a classification.
0 The module consists of:
0 A localization network which takes in the input volume and outputs
parameters of the spatial transformation that should be applied.
0 The creation of a sampling grid that is the result of warping the regular grid
with the affine transformation (theta) created in the localization network.
0 A sampler whose purpose is to perform a warping of the input feature map.
82. Spatial Transformer Network
0 This Network implements the simple idea of making affine
transformations to the input image in order to help models
become more invariant to translation, scale, and rotation.
83. Conclusion
Significant advances in deep reinforcement and
unsupervised learning
Bigger and more complex architectures based on
various interchangeable modules/techniques
Deeper models that can learn from much fewer
training cases
Harder problems such as video understanding
and natural language processing will be
successfully tackled by deep learning algorithms
84. Conclusion
Machines that learn to represent the world from
experience
Deep learning is no magic ! Just statistics in a black
box , but exceptional effective at learning patterns.
We haven’t figured out creativity and human-
empathy.
Transitioning from research to consumer products
.will make the tools you use every day work better,
faster and smarter
#7: AlexNet,8 layers(ILSVRC 2012),VGG,19 layers(ILSVRC 2014),GoogleNet,22 layers(ILSVRC 2014)
ImageNet Large Scale Visual Recognition (ILSVRC )
https://adeshpande3.github.io/adeshpande3.github.io/The-9-Deep-Learning-Papers-You-Need-To-Know-About.html
He, Kaiming, et al. "Deep residual learning for image recognition." Proceedings of the IEEE conference on computer vision and pattern recognition. 2016.
#8: A Beginner's Guide To Understanding Convolutional Neural Networks – Adit Deshpande – CS Undergrad at UCLA ('19)
#9: A Beginner's Guide To Understanding Convolutional Neural Networks – Adit Deshpande – CS Undergrad at UCLA ('19)
#10: A Beginner's Guide To Understanding Convolutional Neural Networks – Adit Deshpande – CS Undergrad at UCLA ('19)
Paw=hand
#13: Convolutional neural networks. Sounds like a weird combination of biology and math with a little CS sprinkled in, but these networks have been some of the most influential innovations in the field of computer vision.
#27: if we computed the network's error for every possible value of a single weight, it would generate the curve you see above. We would then pick the value of the single weight that has the lowest error (the lowest part of the curve). I say single weight because it's a two-dimensional plot. Thus, the x dimension is the value of the weight and the y dimension is the neural network's error when the weight is at that position.
https://iamtrask.github.io/2015/07/27/python-network-part2/
#28: Problem 3: When Slopes are Too Small -Solution 3: Increase the Alpha
#32: Convolutional neural networks(Convnet,CNN)
ConvNet has shown outstanding performance in recognition tasks (image, speech,object)
ConvNet contains hierarchical abstraction process called pooling.
Recurrent neural networks(RNN)
RNN is a generative model: It can generate new data.
Long Short term memory (LSTM) networks[Hochreiter and Schmidhuber, 1997]
RNN+LSTM makes use of long-term memory → Good for time-series data.
Deep/Restricted Boltzmann machines (RBM)
It helps to avoid local minima problem(It regularizes the training data).
But it is not necessary when we have large amount of data.(Drop-out is enough for regularization)
#33: A Beginner's Guide To Understanding Convolutional Neural Networks – Adit Deshpande – CS Undergrad at UCLA ('19)
#34: Every layer of a CNN takes a 3D volume of numbers and outputs a 3D volume of numbers.
#36: However, let’s talk about what this convolution is actually doing from a high level. Each of these filters can be thought of as feature identifiers. When I say features, I’m talking about things like straight edges, simple colors, and curves. Think about the simplest characteristics that all images have in common with each other. Let’s say our first filter is 7 x 7 x 3 and is going to be a curve detector. (In this section, let’s ignore the fact that the filter is 3 units deep and only consider the top depth slice of the filter and the image, for simplicity.)As a curve detector, the filter will have a pixel structure in which there will be higher numerical values along the area that is a shape of a curve (Remember, these filters that we’re talking about as just numbers!).
#39: examples of actual visualizations of the filters of the first conv layer of a trained network. Nonetheless, the main argument remains the same. The filters on the first layer convolve around the input image and “activate” (or compute high values) when the specific feature it is looking for is in the input volume.
#43: https://www.mathworks.com/help/nnet/ug/layers-of-a-convolutional-neural-network.html#mw_ad6f0a9d-9cc7-4e57-9102-0204f1f13e99
https://prateekvjoshi.com/2016/04/05/what-is-local-response-normalization-in-convolutional-neural-networks/
“lateral inhibition”
This refers to the capacity of an excited neuron to subdue its neighbors. We basically want a significant peak so that we have a form of local maxima. This tends to create a contrast in that area, hence increasing the sensory perception. Increasing the sensory perception is a good thing! We want to have the same thing in our CNNs.
#54: Big Data: large data sets are available
– Imagenet: 14M+ labeled images http://www.image-net.org/
– YouTube-8M: 7M+ labeled videos https://research.google.com/youtube8m/
– AWS public data sets: https://aws.amazon.com/public-datasets/
Baidu’s Chinese speech recognition: 4TB of training data, +/- 10 Exaflops
Computations
Neural networks can now be trained on a Raspberry Pi or GPUs because DL model is lightweight
#61: https://adeshpande3.github.io/adeshpande3.github.io/The-9-Deep-Learning-Papers-You-Need-To-Know-About.html
a] "Deep Residual Learning for Image Recognition", Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun. Tech Report 2015. [b] "Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks", Shaoqing Ren, Kaiming He, Ross Girshick, Jian Sun. NIPS 2015.
#62: http://image-net.org/challenges/LSVRC/2016/results
[a] W. Ouyang, X. Wang, X. Zeng, S. Qiu, P. Luo, Y. Tian, H. Li, S. Yang, Z. Wang, C. Loy, X. Tang, “DeepID-Net: Deformable Deep Convolutional Neural Networks for Object Detection,” CVPR 2015. [b] Yang, B., Yan, J., Lei, Z., Li, S. Z. "Craft objects from images." CVPR 2016. [c] X. Zeng, W. Ouyang, B. Yang, J. Yan, X. Wang, “Gated Bi-directional CNN for Object Detection,” ECCV 2016. [d] Ouyang, W., Wang, X., Zhang, C., Yang, X. Factors in Finetuning Deep Model for Object Detection with Long-tail Distribution. CVPR 2016. [e] Wanli Ouyang, Hongyang Li, Xingyu Zeng, and Xiaogang Wang, "Learning Deep Representation with Large-scale Attributes", In Proc. ICCV 2015. [f] https://github.com/yjxiong/caffe
#63: http://image-net.org/challenges/LSVRC/2017/results
[1] Wang F, Jiang M, Qian C, et al. Residual Attention Network for Image Classification[J]. arXiv preprint arXiv:1704.06904, 2017. [2] He K, Zhang X, Ren S, et al. Deep residual learning for image recognition[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2016: 770-778. [3] Szegedy C, Ioffe S, Vanhoucke V, et al. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning[C]//AAAI. 2017: 4278-4284. [4] Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation[J]. arXiv preprint arXiv:1505.04597, 2015. [5] Lin T Y, Dollár P, Girshick R, et al. Feature pyramid networks for object detection[J]. arXiv preprint arXiv:1612.03144, 2016. [6] Shrivastava A, Sukthankar R, Malik J, et al. Beyond skip connections: Top-down modulation for object detection[J]. arXiv preprint arXiv:1612.06851, 2016. [7] Zeng X, Ouyang W, Yan J, et al. Crafting GBD-Net for Object Detection[J]. arXiv preprint arXiv:1610.02579, 2016.
#64: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
#65: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
[N. Dalal and B. Triggs. “Histograms of oriented gradients for human detection.” In CVPR, 2005.]
#66: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
[B. Alexe, T. Deselaers, and V. Ferrari. “Measuring the objectness of image windows.” IEEE transactions on Pattern Analysis and Machine Intelligence, 2012.]
#67: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
#68: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
Texture :بنية او نسيج او تركيب
#69: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
#70: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
[P. F. Felzenszwalb and D. P. Huttenlocher. “Efficient Graph-Based Image Segmentation.” IJCV, 59:167–181, 2004.]
#71: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
[P. F. Felzenszwalb and D. P. Huttenlocher. “Efficient Graph-Based Image Segmentation.” IJCV, 59:167–181, 2004.]
#72: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith
#73: presentation :Selective Search for Object Recognition Uijlings et al. Schuyler Smith