TensorFlow Tutorial given by Dr. Chung-Cheng Chiu at Google Brain on Dec. 29, 2015
http://datasci.tw/event/google_deep_learning
The document provides an introduction to TensorFlow and neural networks. It discusses perceptron classifiers for MNIST data, convolutional neural networks for image recognition, recurrent neural networks for language modeling, and generating poems with RNNs. It also introduces Keras as an easier way to build neural networks and provides contact information for the author and upcoming machine learning conferences.
Video: https://youtu.be/dYhrCUFN0eM
Article: https://medium.com/p/the-gentlest-introduction-to-tensorflow-248dc871a224
Code: https://github.com/nethsix/gentle_tensorflow/blob/master/code/linear_regression_one_feature.py
This alternative introduction to Google's official Tensorflow (TF) tutorial strips away the unnecessary concepts that overly complicates getting started. The goal is to use TF to perform Linear Regression (LR) that has only a single-feature. We show how to model the LR using a TF graph, how to define the cost function to measure how well the an LR model fits the dataset, and finally train the LR model to find the best fit model.
Gentlest Introduction to Tensorflow - Part 3Khor SoonHin
Articles:
* https://medium.com/all-of-us-are-belong-to-machines/gentlest-intro-to-tensorflow-part-3-matrices-multi-feature-linear-regression-30a81ebaaa6c
* https://medium.com/all-of-us-are-belong-to-machines/gentlest-intro-to-tensorflow-4-logistic-regression-2afd0cabc54
Video: https://youtu.be/F8g_6TXKlxw
Code: https://github.com/nethsix/gentle_tensorflow
In this part, we:
* Use Tensorflow for linear regression models with multiple features
* Use Tensorflow for logistic regression models with multiple features. Specifically:
* Predict multi-class/discrete outcome
* Explain why we use cross-entropy as cost function
* Explain why we use softmax
* Tensorflow Cheatsheet #1
* Single feature linear regression
* Multi-feature linear regression
* Multi-feature logistic regression
The document contains code snippets demonstrating the use of TensorFlow for building and training neural networks. It shows how to:
1. Define operations like convolutions, max pooling, fully connected layers using common TensorFlow functions like tf.nn.conv2d and tf.nn.max_pool.
2. Create and initialize variables using tf.Variable and initialize them using tf.global_variables_initializer.
3. Construct a multi-layer perceptron model for MNIST classification with convolutional and fully connected layers.
4. Train the model using tf.train.AdamOptimizer by running optimization steps and calculating loss over batches of data.
5. Evaluate the trained model on test data to calculate accuracy.
Explanation on Tensorflow example -Deep mnist for expert홍배 김
you can find the exact and detailed network architecture of 'Deep mnist for expert' example of tensorflow's tutorial. I also added descriptions on the program for your better understanding.
An introduction to Google's AI Engine, look deeper into Artificial Networks and Machine Learning. Appreciate how our simplest neural network be codified and be used to data analytics.
TensorFlow is an open source neural network library for Python and C++. It defines data flows as graphs with nodes representing operations and edges representing multidimensional data arrays called tensors. It supports supervised learning algorithms like gradient descent to minimize cost functions. TensorFlow automatically computes gradients so the user only needs to define the network structure, cost function, and optimization algorithm. An example shows training various neural network models on the MNIST handwritten digit dataset, achieving up to 99.2% accuracy. TensorFlow can implement other models like recurrent neural networks and is a simple yet powerful framework for neural networks.
TensorFlow is a wonderful tool for rapidly implementing neural networks. In this presentation, we will learn the basics of TensorFlow and show how neural networks can be built with just a few lines of code. We will highlight some of the confusing bits of TensorFlow as a way of developing the intuition necessary to avoid common pitfalls when developing your own models. Additionally, we will discuss how to roll our own Recurrent Neural Networks. While many tutorials focus on using built in modules, this presentation will focus on writing neural networks from scratch enabling us to build flexible models when Tensorflow’s high level components can’t quite fit our needs.
About Nathan Lintz:
Nathan Lintz is a research scientist at indico Data Solutions where he is responsible for developing machine learning systems in the domains of language detection, text summarization, and emotion recognition. Outside of work, Nathan is currently writting a book on TensorFlow as an extension to his tutorial repository https://github.com/nlintz/TensorFlow-Tutorials
Link to video https://www.youtube.com/watch?v=op1QJbC2g0E&feature=youtu.be
The document describes how to use TensorBoard, TensorFlow's visualization tool. It outlines 5 steps: 1) annotate nodes in the TensorFlow graph to visualize, 2) merge summaries, 3) create a writer, 4) run the merged summary and write it, 5) launch TensorBoard pointing to the log directory. TensorBoard can visualize the TensorFlow graph, plot metrics over time, and show additional data like histograms and scalars.
Gentlest Introduction to Tensorflow - Part 2Khor SoonHin
Video: https://youtu.be/Trc52FvMLEg
Article: https://medium.com/@khor/gentlest-introduction-to-tensorflow-part-2-ed2a0a7a624f
Code: https://github.com/nethsix/gentle_tensorflow
Continuing from Part 1 where we used Tensorflow to perform linear regression for a model with single feature, here we:
* Use Tensorboard to visualize linear regression variables and the Tensorflow network graph
* Perform stochastic/mini-batch/batch gradient descent
Abstract: This PDSG workshop introduces basic concepts on TensorFlow. The course covers fundamentals. Concepts covered are Vectors/Matrices/Vectors, Design&Run, Constants, Operations, Placeholders, Bindings, Operators, Loss Function and Training.
Level: Fundamental
Requirements: Some basic programming knowledge is preferred. No prior statistics background is required.
This document provides a summary of a 30-minute presentation on feature selection in Python. The presentation covered several common feature selection techniques in Python like LASSO, random forests, and PCA. Code examples were provided to demonstrate how to perform feature selection on the Iris dataset using these techniques in scikit-learn. Dimensionality reduction with PCA and word embeddings with Gensim were also briefly discussed. The presentation aimed to provide practical code examples to do feature selection without explanations of underlying mathematics or theory.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs, followed by a Keras code sample for defining a CNN. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session. Then we'll see a short introduction to TensorFlow 1.x and some insights into TF 2 that will be released some time this year.
A fast-paced introduction to TensorFlow 2 about some important new features (such as generators and the @tf.function decorator) and TF 1.x functionality that's been removed from TF 2 (yes, tf.Session() has retired).
Concise code samples are presented to illustrate how to use new features of TensorFlow 2. You'll also get a quick introduction to lazy operators (if you know FRP this will be super easy), along with a code comparison between TF 1.x/iterators with tf.data.Dataset and TF 2/generators with tf.data.Dataset.
Finally, we'll look at some tf.keras code samples that are based on TensorFlow 2. Although familiarity with TF 1.x is helpful, newcomers with an avid interest in learning about TensorFlow 2 can benefit from this session.
This session for beginners introduces tf.data APIs for creating data pipelines by combining various "lazy operators" in tf.data, such as filter(), map(), batch(), zip(), flatmap(), take(), and so forth.
Familiarity with method chaining and TF2 is helpful (but not required). If you are comfortable with FRP, the code samples in this session will be very familiar to you.
Introduction to Deep Learning, Keras, and TensorflowOswald Campesato
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session. Then we'll see how to create a Convolutional Neural Network in Keras, followed by a quick introduction to TensorFlow and TensorBoard.
A fast-paced introduction to TensorFlow 2 about some important new features (such as generators and the @tf.function decorator) and TF 1.x functionality that's been removed from TF 2 (yes, tf.Session() has retired).
Some concise code samples are presented to illustrate how to use new features of TensorFlow 2.
This document provides an overview and introduction to TensorFlow 2. It discusses major changes from TensorFlow 1.x like eager execution and tf.function decorator. It covers working with tensors, arrays, datasets, and loops in TensorFlow 2. It also demonstrates common operations like arithmetic, reshaping and normalization. Finally, it briefly introduces working with Keras and neural networks in TensorFlow 2.
An introductory presentation covered key concepts in deep learning including neural networks, activation functions, cost functions, and optimization methods. Popular deep learning frameworks TensorFlow and tensorflow.js were discussed. Common deep learning architectures like convolutional neural networks and generative adversarial networks were explained. Examples and code snippets in Python demonstrated fundamental deep learning concepts.
This document provides an overview of TensorFlow and how to implement machine learning models using TensorFlow. It discusses:
1) How to install TensorFlow either directly or within a virtual environment.
2) The key concepts of TensorFlow including computational graphs, sessions, placeholders, variables and how they are used to define and run computations.
3) An example one-layer perceptron model for MNIST image classification to demonstrate these concepts in action.
This book is intended for education and fun. Python is an amazing, text-based coding language, perfectly suited for children older than the age of 10. The Standard Python library has a module called Turtle which is a popular way to introduce programming to kids. This library enables children to create pictures and shapes by providing them with a virtual canvas. With the Python Turtle library, you can create nice animation projects using images that are taken from the internet, scaled-down stored as a gif-files download to the projects. The book includes 19 basic lessons with examples that introduce to the Python codes through Turtle library which is convenient to the school students of 10+years old. The book has also a lot of projects that show how to make different animations with Turtle graphics: games, applications to math, physics, and science.
This document discusses computer vision applications using TensorFlow for deep learning. It introduces computer vision and convolutional neural networks. It then demonstrates how to build and train a CNN for MNIST handwritten digit recognition using TensorFlow. Finally, it shows how to load and run the pre-trained Google Inception model for image classification.
This document provides an introduction to GUI programming using Tkinter and turtle graphics in Python. It discusses how turtle programs use Tkinter to create windows and display graphics. Tkinter is a Python package for building graphical user interfaces based on the Tk widget toolkit. Several examples are provided on how to use turtle graphics to draw shapes, add color and interactivity using keyboard and mouse inputs. GUI programming with Tkinter allows creating more complex programs and games beyond what can be done with turtle graphics alone.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session. Then we'll see a short introduction to TensorFlow and TensorBoard.
Distilled PyTorch tutorial. Also in text at my blog - https://towardsdatascience.com/pytorch-tutorial-distilled-95ce8781a89c
RNNs are neural networks that can handle sequence data by incorporating a time component. They learn from past sequence data to predict future states in new sequence data. The document discusses RNN architecture, which includes an input layer, hidden layer, and output layer. The hidden layer receives input from the current time step and previous hidden state. Backpropagation Through Time is used to train RNNs by propagating error terms back in time. The document provides an example implementation of an RNN for time series prediction using TensorFlow and Keras.
RNNs are neural networks that can handle sequence data by incorporating a time component. They learn from past sequence data to predict future states in new sequence data. The document discusses RNN architecture, which uses a hidden layer that receives both the current input and the previous hidden state. It also covers backpropagation through time (BPTT) for training RNNs on sequence data. Examples are provided to implement an RNN from scratch using TensorFlow and Keras to predict a noisy sine wave time series.
TensorFlow is an open source software library for machine learning developed by Google. It provides primitives for defining functions on tensors and automatically computing their derivatives. TensorFlow represents computations as data flow graphs with nodes representing operations and edges representing tensors. It is widely used for neural networks and deep learning tasks like image classification, language processing, and speech recognition. TensorFlow is portable, scalable, and has a large community and support for deployment compared to other frameworks. It works by constructing a computational graph during modeling, and then executing operations by pushing data through the graph.
Large Scale Deep Learning with TensorFlow Jen Aman
Large-scale deep learning with TensorFlow allows storing and performing computation on large datasets to develop computer systems that can understand data. Deep learning models like neural networks are loosely based on what is known about the brain and become more powerful with more data, larger models, and more computation. At Google, deep learning is being applied across many products and areas, from speech recognition to image understanding to machine translation. TensorFlow provides an open-source software library for machine learning that has been widely adopted both internally at Google and externally.
The document describes how to use TensorBoard, TensorFlow's visualization tool. It outlines 5 steps: 1) annotate nodes in the TensorFlow graph to visualize, 2) merge summaries, 3) create a writer, 4) run the merged summary and write it, 5) launch TensorBoard pointing to the log directory. TensorBoard can visualize the TensorFlow graph, plot metrics over time, and show additional data like histograms and scalars.
Gentlest Introduction to Tensorflow - Part 2Khor SoonHin
Video: https://youtu.be/Trc52FvMLEg
Article: https://medium.com/@khor/gentlest-introduction-to-tensorflow-part-2-ed2a0a7a624f
Code: https://github.com/nethsix/gentle_tensorflow
Continuing from Part 1 where we used Tensorflow to perform linear regression for a model with single feature, here we:
* Use Tensorboard to visualize linear regression variables and the Tensorflow network graph
* Perform stochastic/mini-batch/batch gradient descent
Abstract: This PDSG workshop introduces basic concepts on TensorFlow. The course covers fundamentals. Concepts covered are Vectors/Matrices/Vectors, Design&Run, Constants, Operations, Placeholders, Bindings, Operators, Loss Function and Training.
Level: Fundamental
Requirements: Some basic programming knowledge is preferred. No prior statistics background is required.
This document provides a summary of a 30-minute presentation on feature selection in Python. The presentation covered several common feature selection techniques in Python like LASSO, random forests, and PCA. Code examples were provided to demonstrate how to perform feature selection on the Iris dataset using these techniques in scikit-learn. Dimensionality reduction with PCA and word embeddings with Gensim were also briefly discussed. The presentation aimed to provide practical code examples to do feature selection without explanations of underlying mathematics or theory.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs, followed by a Keras code sample for defining a CNN. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session. Then we'll see a short introduction to TensorFlow 1.x and some insights into TF 2 that will be released some time this year.
A fast-paced introduction to TensorFlow 2 about some important new features (such as generators and the @tf.function decorator) and TF 1.x functionality that's been removed from TF 2 (yes, tf.Session() has retired).
Concise code samples are presented to illustrate how to use new features of TensorFlow 2. You'll also get a quick introduction to lazy operators (if you know FRP this will be super easy), along with a code comparison between TF 1.x/iterators with tf.data.Dataset and TF 2/generators with tf.data.Dataset.
Finally, we'll look at some tf.keras code samples that are based on TensorFlow 2. Although familiarity with TF 1.x is helpful, newcomers with an avid interest in learning about TensorFlow 2 can benefit from this session.
This session for beginners introduces tf.data APIs for creating data pipelines by combining various "lazy operators" in tf.data, such as filter(), map(), batch(), zip(), flatmap(), take(), and so forth.
Familiarity with method chaining and TF2 is helpful (but not required). If you are comfortable with FRP, the code samples in this session will be very familiar to you.
Introduction to Deep Learning, Keras, and TensorflowOswald Campesato
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session. Then we'll see how to create a Convolutional Neural Network in Keras, followed by a quick introduction to TensorFlow and TensorBoard.
A fast-paced introduction to TensorFlow 2 about some important new features (such as generators and the @tf.function decorator) and TF 1.x functionality that's been removed from TF 2 (yes, tf.Session() has retired).
Some concise code samples are presented to illustrate how to use new features of TensorFlow 2.
This document provides an overview and introduction to TensorFlow 2. It discusses major changes from TensorFlow 1.x like eager execution and tf.function decorator. It covers working with tensors, arrays, datasets, and loops in TensorFlow 2. It also demonstrates common operations like arithmetic, reshaping and normalization. Finally, it briefly introduces working with Keras and neural networks in TensorFlow 2.
An introductory presentation covered key concepts in deep learning including neural networks, activation functions, cost functions, and optimization methods. Popular deep learning frameworks TensorFlow and tensorflow.js were discussed. Common deep learning architectures like convolutional neural networks and generative adversarial networks were explained. Examples and code snippets in Python demonstrated fundamental deep learning concepts.
This document provides an overview of TensorFlow and how to implement machine learning models using TensorFlow. It discusses:
1) How to install TensorFlow either directly or within a virtual environment.
2) The key concepts of TensorFlow including computational graphs, sessions, placeholders, variables and how they are used to define and run computations.
3) An example one-layer perceptron model for MNIST image classification to demonstrate these concepts in action.
This book is intended for education and fun. Python is an amazing, text-based coding language, perfectly suited for children older than the age of 10. The Standard Python library has a module called Turtle which is a popular way to introduce programming to kids. This library enables children to create pictures and shapes by providing them with a virtual canvas. With the Python Turtle library, you can create nice animation projects using images that are taken from the internet, scaled-down stored as a gif-files download to the projects. The book includes 19 basic lessons with examples that introduce to the Python codes through Turtle library which is convenient to the school students of 10+years old. The book has also a lot of projects that show how to make different animations with Turtle graphics: games, applications to math, physics, and science.
This document discusses computer vision applications using TensorFlow for deep learning. It introduces computer vision and convolutional neural networks. It then demonstrates how to build and train a CNN for MNIST handwritten digit recognition using TensorFlow. Finally, it shows how to load and run the pre-trained Google Inception model for image classification.
This document provides an introduction to GUI programming using Tkinter and turtle graphics in Python. It discusses how turtle programs use Tkinter to create windows and display graphics. Tkinter is a Python package for building graphical user interfaces based on the Tk widget toolkit. Several examples are provided on how to use turtle graphics to draw shapes, add color and interactivity using keyboard and mouse inputs. GUI programming with Tkinter allows creating more complex programs and games beyond what can be done with turtle graphics alone.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session. Then we'll see a short introduction to TensorFlow and TensorBoard.
Distilled PyTorch tutorial. Also in text at my blog - https://towardsdatascience.com/pytorch-tutorial-distilled-95ce8781a89c
RNNs are neural networks that can handle sequence data by incorporating a time component. They learn from past sequence data to predict future states in new sequence data. The document discusses RNN architecture, which includes an input layer, hidden layer, and output layer. The hidden layer receives input from the current time step and previous hidden state. Backpropagation Through Time is used to train RNNs by propagating error terms back in time. The document provides an example implementation of an RNN for time series prediction using TensorFlow and Keras.
RNNs are neural networks that can handle sequence data by incorporating a time component. They learn from past sequence data to predict future states in new sequence data. The document discusses RNN architecture, which uses a hidden layer that receives both the current input and the previous hidden state. It also covers backpropagation through time (BPTT) for training RNNs on sequence data. Examples are provided to implement an RNN from scratch using TensorFlow and Keras to predict a noisy sine wave time series.
TensorFlow is an open source software library for machine learning developed by Google. It provides primitives for defining functions on tensors and automatically computing their derivatives. TensorFlow represents computations as data flow graphs with nodes representing operations and edges representing tensors. It is widely used for neural networks and deep learning tasks like image classification, language processing, and speech recognition. TensorFlow is portable, scalable, and has a large community and support for deployment compared to other frameworks. It works by constructing a computational graph during modeling, and then executing operations by pushing data through the graph.
Large Scale Deep Learning with TensorFlow Jen Aman
Large-scale deep learning with TensorFlow allows storing and performing computation on large datasets to develop computer systems that can understand data. Deep learning models like neural networks are loosely based on what is known about the brain and become more powerful with more data, larger models, and more computation. At Google, deep learning is being applied across many products and areas, from speech recognition to image understanding to machine translation. TensorFlow provides an open-source software library for machine learning that has been widely adopted both internally at Google and externally.
生成式對抗網路 (Generative Adversarial Network, GAN) 顯然是深度學習領域的下一個熱點,Yann LeCun 說這是機器學習領域這十年來最有趣的想法 (the most interesting idea in the last 10 years in ML),又說這是有史以來最酷的東西 (the coolest thing since sliced bread)。生成式對抗網路解決了什麼樣的問題呢?在機器學習領域,回歸 (regression) 和分類 (classification) 這兩項任務的解法人們已經不再陌生,但是如何讓機器更進一步創造出有結構的複雜物件 (例如:圖片、文句) 仍是一大挑戰。用生成式對抗網路,機器已經可以畫出以假亂真的人臉,也可以根據一段敘述文字,自己畫出對應的圖案,甚至還可以畫出二次元人物頭像 (左邊的動畫人物頭像就是機器自己生成的)。本課程希望能帶大家認識生成式對抗網路這個深度學習最前沿的技術。
- TensorFlow is Google's open source machine learning library for developing and training neural networks and deep learning models. It operates using data flow graphs to represent computation.
- TensorFlow can be used across many platforms including data centers, CPUs, GPUs, mobile phones, and IoT devices. It is widely used at Google across many products and research areas involving machine learning.
- The TensorFlow library is used along with higher level tools in Google's machine learning platform including TensorFlow Cloud, Machine Learning APIs, and Cloud Machine Learning Platform to make machine learning more accessible and scalable.
TensorFlow Serving, Deep Learning on Mobile, and Deeplearning4j on the JVM - ...Sam Putnam [Deep Learning]
1) The document discusses TensorFlow Serving, Deep Learning on Mobile, and Deeplearning4j on the JVM as presented by Sam Putnam on 6/8/2017.
2) It provides information on exporting models for TensorFlow Serving, deploying TensorFlow to Android, tools for mobile deep learning like Inception and MobileNets, and using Deeplearning4j on the JVM with integration with Spark.
3) The document shares links to resources on these topics and thanks sponsors while inviting people to join future discussions.
This slides explains how Convolution Neural Networks can be coded using Google TensorFlow.
Video available at : https://www.youtube.com/watch?v=EoysuTMmmMc
Introducing TensorFlow: The game changer in building "intelligent" applicationsRokesh Jankie
This is the slidedeck used for the presentation of the Amsterdam Pipeline of Data Science, held in December 2016. TensorFlow in the open source library from Google to implement deep learning, neural networks. This is an introduction to Tensorflow.
Note: Videos are not included (which were shown during the presentation)
On-device machine learning: TensorFlow on AndroidYufeng Guo
This document discusses building machine learning models for mobile apps using TensorFlow. It describes the process of gathering training data, training a model using Cloud ML Engine, optimizing the model for mobile, and integrating it into an Android app. Key steps involve converting video training data to images, retraining an InceptionV3 model, optimizing the model size with graph transformations, and loading the model into an Android app. TensorFlow allows developing machine learning models that can run efficiently on mobile devices.
Deep Learning for Data Scientists - Data Science ATL Meetup Presentation, 201...Andrew Gardner
Note: these are the slides from a presentation at Lexis Nexis in Alpharetta, GA, on 2014-01-08 as part of the DataScienceATL Meetup. A video of this talk from Dec 2013 is available on vimeo at http://bit.ly/1aJ6xlt
Note: Slideshare mis-converted the images in slides 16-17. Expect a fix in the next couple of days.
---
Deep learning is a hot area of machine learning named one of the "Breakthrough Technologies of 2013" by MIT Technology Review. The basic ideas extend neural network research from past decades and incorporate new discoveries in statistical machine learning and neuroscience. The results are new learning architectures and algorithms that promise disruptive advances in automatic feature engineering, pattern discovery, data modeling and artificial intelligence. Empirical results from real world applications and benchmarking routinely demonstrate state-of-the-art performance across diverse problems including: speech recognition, object detection, image understanding and machine translation. The technology is employed commercially today, notably in many popular Google products such as Street View, Google+ Image Search and Android Voice Recognition.
In this talk, we will present an overview of deep learning for data scientists: what it is, how it works, what it can do, and why it is important. We will review several real world applications and discuss some of the key hurdles to mainstream adoption. We will conclude by discussing our experiences implementing and running deep learning experiments on our own hardware data science appliance.
Machine Learning Preliminaries and Math Refresherbutest
The document is an introduction to machine learning preliminaries and mathematics. It covers general remarks about learning as a process of model building, an overview of key concepts from probability theory and statistics needed for machine learning like random variables, distributions, and expectations. It also introduces linear spaces and vector spaces as mathematical structures that are important foundations for machine learning algorithms. The goal is to cover essential mathematical concepts like probability, statistics, and linear algebra that are prerequisites for machine learning.
Secure Because Math: A Deep-Dive on Machine Learning-Based Monitoring (#Secur...Alex Pinto
The document discusses machine learning-based security monitoring. It begins with an introduction of the speaker, Alex Pinto, and an agenda that will include a discussion of anomaly detection versus classification techniques. It then covers some history of anomaly detection research dating back to the 1980s. It also discusses challenges with anomaly detection, such as the curse of dimensionality with high-dimensional data and lack of ground truth labels. The document emphasizes communicating these machine learning concepts clearly.
Machine Learning without the Math: An overview of Machine LearningArshad Ahmed
A brief overview of Machine Learning and its associated tasks from a high level. This presentation discusses key concepts without the maths.The more mathematically inclined are referred to Bishops book on Pattern Recognition and Machine Learning.
qconsf 2013: Top 10 Performance Gotchas for scaling in-memory Algorithms - Sr...Sri Ambati
Top 10 Performance Gotchas in scaling in-memory Algorithms
Abstract:
Math Algorithms have primarily been the domain of desktop data science. With the success of scalable algorithms at Google, Amazon, and Netflix, there is an ever growing demand for sophisticated algorithms over big data. In this talk, we get a ringside view in the making of the world's most scalable and fastest machine learning framework, H2O, and the performance lessons learnt scaling it over EC2 for Netflix and over commodity hardware for other power users.
Top 10 Performance Gotchas is about the white hot stories of i/o wars, S3 resets, and muxers, as well as the power of primitive byte arrays, non-blocking structures, and fork/join queues. Of good data distribution & fine-grain decomposition of Algorithms to fine-grain blocks of parallel computation. It's a 10-point story of the rage of a network of machines against the tyranny of Amdahl while keeping the statistical properties of the data and accuracy of the algorithm.
Track: Scalability, Availability, and Performance: Putting It All Together
Time: Wednesday, 11:45am - 12:35pm
1) Machine learning draws on areas of mathematics including probability, statistical inference, linear algebra, and optimization theory.
2) While there are easy-to-use machine learning packages, understanding the underlying mathematics is important for choosing the right algorithms, making good parameter and validation choices, and interpreting results.
3) Key concepts in probability and statistics that are important for machine learning include random variables, probability distributions, expected value, variance, covariance, and conditional probability. These concepts allow quantification of relationships and uncertainties in data.
Kafka Summit SF Apr 26 2016 - Generating Real-time Recommendations with NiFi,...Chris Fregly
This document summarizes a presentation about generating real-time streaming recommendations using NiFi, Kafka, and Spark ML. The presentation demonstrates using NiFi to ingest data from HTTP requests, enrich it with geo data, and write it to a Kafka topic. It then shows how to create a Spark Streaming application that reads from Kafka to perform incremental matrix factorization recommendations in real-time and handles failures using circuit breakers. The presentation also provides an overview of Netflix's large-scale real-time recommendation pipeline.
Big Data Spain - Nov 17 2016 - Madrid Continuously Deploy Spark ML and Tensor...Chris Fregly
In this talk, I describe some recent advancements in Streaming ML and AI Pipelines to enable data scientists to rapidly train and test on streaming data - and ultimately deploy models directly into production on their own with low friction and high impact.
With proper tooling and monitoring, data scientist have the freedom and responsibility to experiment rapidly on live, streaming data - and deploy directly into production as often as necessary. I’ll describe this tooling - and demonstrate a real production pipeline using Jupyter Notebook, Docker, Kubernetes, Spark ML, Kafka, TensorFlow, Jenkins, and Netflix Open Source.
Workshop about TensorFlow usage for AI Ukraine 2016. Brief tutorial with source code example. Described TensorFlow main ideas, terms, parameters. Example related with linear neuron model and learning using Adam optimization algorithm.
This document provides an introduction and overview of TensorFlow, a popular deep learning library developed by Google. It begins with administrative announcements for the class and then discusses key TensorFlow concepts like tensors, variables, placeholders, sessions, and computation graphs. It provides examples comparing TensorFlow and NumPy for common deep learning tasks like linear regression. It also covers best practices for debugging TensorFlow and introduces TensorBoard for visualization. Overall, the document serves as a high-level tutorial for getting started with TensorFlow.
Tensor flow description of ML Lab. documentjeongok1
This document contains slides for a TensorFlow basics lab. It introduces TensorFlow and computational graphs, shows how to install TensorFlow and check the version, and demonstrates a simple "Hello World" TensorFlow program. It also discusses placeholders, variables, feeding data, and building linear regression models in TensorFlow to minimize a cost function. The full Python code for linear regression is provided.
A fast-paced introduction to Deep Learning (DL) concepts, starting with a simple yet complete neural network (no frameworks), followed by aspects of deep neural networks, such as back propagation, activation functions, CNNs, and the AUT theorem. Next, a quick introduction to TensorFlow and TensorBoard, along with some code samples with TensorFlow. For best results, familiarity with basic vectors and matrices, inner (aka "dot") products of vectors, the notion of a derivative, and rudimentary Python is recommended.
Introduction to Deep Learning, Keras, and TensorFlowSri Ambati
This meetup was recorded in San Francisco on Jan 9, 2019.
Video recording of the session can be viewed here: https://youtu.be/yG1UJEzpJ64
Description:
This fast-paced session starts with a simple yet complete neural network (no frameworks), followed by an overview of activation functions, cost functions, backpropagation, and then a quick dive into CNNs. Next, we'll create a neural network using Keras, followed by an introduction to TensorFlow and TensorBoard. For best results, familiarity with basic vectors and matrices, inner (aka "dot") products of vectors, and rudimentary Python is definitely helpful. If time permits, we'll look at the UAT, CLT, and the Fixed Point Theorem. (Bonus points if you know Zorn's Lemma, the Well-Ordering Theorem, and the Axiom of Choice.)
Oswald's Bio:
Oswald Campesato is an education junkie: a former Ph.D. Candidate in Mathematics (ABD), with multiple Master's and 2 Bachelor's degrees. In a previous career, he worked in South America, Italy, and the French Riviera, which enabled him to travel to 70 countries throughout the world.
He has worked in American and Japanese corporations and start-ups, as C/C++ and Java developer to CTO. He works in the web and mobile space, conducts training sessions in Android, Java, Angular 2, and ReactJS, and he writes graphics code for fun. He's comfortable in four languages and aspires to become proficient in Japanese, ideally sometime in the next two decades. He enjoys collaborating with people who share his passion for learning the latest cool stuff, and he's currently working on his 15th book, which is about Angular 2.
Intro to Deep Learning, TensorFlow, and tensorflow.jsOswald Campesato
This fast-paced session introduces Deep Learning concepts, such gradient descent, back propagation, activation functions, and CNNs. We'll look at creating Android apps with TensorFlow Lite (pending its availability). Basic knowledge of vectors, matrices, and Android, as well as elementary calculus (derivatives), are strongly recommended in order to derive the maximum benefit from this session.
An introductory document covered deep learning concepts including neural networks, activation functions, cost functions, gradient descent, TensorFlow, CNNs, RNNs, GANs, and tensorflow.js. Key topics included the use of deep learning for computer vision, speech recognition, and more. Activation functions such as ReLU, sigmoid and tanh were explained. TensorFlow and tensorflow.js were introduced as frameworks for deep learning.
Lucio Floretta - TensorFlow and Deep Learning without a PhD - Codemotion Mila...Codemotion
With TensorFlow, deep machine learning transitions from an area of research to mainstream software engineering. In this session, we'll work together to construct and train a neural network that recognises handwritten digits. Along the way, we'll discover some of the "tricks of the trade" used in neural network design, and finally, we'll bring the recognition accuracy of our model above 99%.
This document provides an overview of TensorFlow presented by Ashish Agarwal and Ashish Bansal. The key points covered include:
- TensorFlow is an open-source machine learning framework for research and production. It allows models to be deployed across different platforms.
- TensorFlow models are represented as dataflow graphs where nodes are operations and edges are tensors flowing between operations. Placeholders, variables, and sessions are introduced.
- Examples demonstrate basic linear regression and logistic regression models built with TensorFlow. Layers API and common neural network components like convolutions and RNNs are also covered.
- Advanced models like AlexNet, Inception, ResNet, and neural machine translation with attention are briefly overviewed.
This document summarizes TensorFlow's APIs, beginning with an overview of the low-level API using computational graphs and sessions. It then discusses higher-level APIs like Keras, TensorFlow Datasets for input pipelines, and Estimators which hide graph and session details. Datasets improve training speed by up to 300% by enabling parallelism. Estimators resemble scikit-learn and separate model definition from training, making code more modular and reusable. The document provides examples of using Datasets and Estimators with TensorFlow.
This fast-paced session starts with an introduction to neural networks and linear regression models, along with a quick view of TensorFlow, followed by some Scala APIs for TensorFlow. You'll also see a simple dockerized image of Scala and TensorFlow code and how to execute the code in that image from the command line. No prior knowledge of NNs, Keras, or TensorFlow is required (but you must be comfortable with Scala).
Introduction To TensorFlow | Deep Learning Using TensorFlow | CloudxLabCloudxLab
This document provides instructions for getting started with TensorFlow using a free CloudxLab. It outlines the following steps:
1. Open CloudxLab and enroll if not already enrolled. Otherwise go to "My Lab".
2. In "My Lab", open Jupyter and run commands to clone an ML repository containing TensorFlow examples.
3. Go to the deep learning folder in Jupyter and open the TensorFlow notebook to get started with examples.
The release of TensorFlow 2.0 comes with a significant number of improvements over its 1.x version, all with a focus on ease of usability and a better user experience. We will give an overview of what TensorFlow 2.0 is and discuss how to get started building models from scratch using TensorFlow 2.0’s high-level api, Keras. We will walk through an example step-by-step in Python of how to build an image classifier. We will then showcase how to leverage a transfer learning to make building a model even easier! With transfer learning, we can leverage other pretrained models such as ImageNet to drastically speed up the training time of our model. TensorFlow 2.0 makes this incredibly simple to do.
Introduction To Using TensorFlow & Deep Learningali alemi
This document provides an introduction to using TensorFlow. It begins with an overview of TensorFlow and what it is. It then discusses TensorFlow code basics, including building computational graphs and running sessions. It provides examples of using placeholders, constants, and variables. It also gives an example of linear regression using TensorFlow. Finally, it discusses deep learning techniques like convolutional neural networks (CNNs) and recurrent neural networks (RNNs), providing examples of CNNs for image classification. It concludes with an example of using a multi-layer perceptron for MNIST digit classification in TensorFlow.
The document introduces TensorFlow, a machine learning library. It discusses how TensorFlow uses multi-dimensional arrays called tensors to represent data and models. An example regression problem is demonstrated where TensorFlow is used to fit a line to sample data points by iteratively updating the slope and offset values to minimize loss. The document promotes TensorFlow by noting its ability to distribute operations across processors and optimize entire graphs.
Introduction to TensorFlow, by Machine Learning at BerkeleyTed Xiao
A workshop introducing the TensorFlow Machine Learning framework. Presented by Brenton Chu, Vice President of Machine Learning at Berkeley.
This presentation cover show to construct, train, evaluate, and visualize neural networks in TensorFlow 1.0
http://ml.berkeley.edu
TensorFlow and Keras are popular deep learning frameworks. TensorFlow is an open source library for numerical computation using data flow graphs. It was developed by Google and is widely used for machine learning and deep learning. Keras is a higher-level neural network API that can run on top of TensorFlow. It focuses on user-friendliness, modularization and extensibility. Both frameworks make building and training neural networks easier through modular layers and built-in optimization algorithms.
In this article you will learn hot to use tensorflow Softmax Classifier estimator to classify MNIST dataset in one script.
This paper introduces also the basic idea of a artificial neural network.
Language translation with Deep Learning (RNN) with TensorFlowS N
This document provides an overview of a meetup on language translation with deep learning using TensorFlow on FloydHub. It will cover the language translation challenge, introducing key concepts like deep learning, RNNs, NLP, TensorFlow and FloydHub. It will then describe the solution approach to the translation task, including a demo and code walkthrough. Potential next steps and references for further learning are also mentioned.
Машинное обучение на JS. С чего начать и куда идти | Odessa Frontend Meetup #12OdessaFrontend
Tensorflow.js allows developing machine learning models with JavaScript. It provides APIs for building neural networks with layers, training models on data, and making predictions. Key capabilities include building, training, and deploying deep learning models directly in the browser or on the server using JavaScript. It supports common ML tasks like image classification, object detection, and natural language processing.
This document is a presentation by Ted Chang about creating new opportunities for Taiwan's intelligent transformation. It discusses paradigm shifts in technology such as mobile phones and cloud computing. It introduces concepts like the Internet of Things, artificial intelligence, and how they can be combined. It argues that key driving forces for the future will be machine learning, big data, cloud computing and AI. The presentation envisions applications of these technologies in areas like future medicine and smart manufacturing. It ends by emphasizing the importance of wisdom and intelligence in shaping the future.
- The document discusses how artificial intelligence can enable earlier and safer medicine.
- It provides background on the author and their expertise in biomedical informatics and roles as editor-in-chief of several academic journals.
- Key applications of AI in healthcare discussed include using machine learning on large medical datasets to detect suspicious moles earlier, reduce medication errors, and more accurately predict cancer occurrence up to 12 months in advance.
- The author argues that AI has the potential to transform medicine by enabling more preventive and earlier detection approaches compared to traditional reactive healthcare models.
Jane may be able to help. Let me check with her personal assistant Jane-ML.
NextPrevIndex
Meera checks with Jane-ML
User-Agent Interaction (V)
48
PA_Meera: Mina, do you
have trouble in
debugging?
Mina: Yes, is there
anyone who has done
this?
Personal Agent
[Meera]
Jane-ML: Jane has done a similar debugging problem before. She is available now and willing to help.
compiletheme
Compiling output
1) Kaggle is the largest platform for AI and data science competitions, acquired by Google in 2017. It has been used by companies like Bosch, Mercedes, and Asus for challenges like improving production lines, accelerating testing processes, and component failure prediction.
2) The document discusses the author's experiences winning silver medals in Kaggle competitions involving camera model identification, passenger screening algorithms, and pneumonia detection. For camera model identification, the author used transfer learning with InceptionResNetV2 and high-pass filters to identify camera models from images.
3) For passenger screening, the author modified a 2D CNN to 3D and used 3D data augmentation to rank in the top 7% of the $1
[台灣人工智慧學校] Bridging AI to Precision Agriculture through IoT台灣資料科學年會
The document describes a system for precision agriculture using IoT. It involves sensors collecting environmental data from fields and feeding it to a control board connected to actuators like irrigation systems. The data is also sent to an IoTtalk engine and AgriTalk server in the cloud for analysis and remote access/control through an AgriGUI interface. Equations were developed to estimate nutrient levels like nitrogen from sensor readings to help optimize crop growth.
The document discusses Open Robot Club and includes several links to its website and YouTube videos. It provides information on the club's computing resources like NVIDIA V100 GPUs. Tables with metrics like underkill and overkill percentages are included for different types of tasks like AI AOI and PCB inspection. The club's website and demos are referenced throughout.
European Accessibility Act & Integrated Accessibility TestingJulia Undeutsch
Emma Dawson will guide you through two important topics in this session.
Firstly, she will prepare you for the European Accessibility Act (EAA), which comes into effect on 28 June 2025, and show you how development teams can prepare for it.
In the second part of the webinar, Emma Dawson will explore with you various integrated testing methods and tools that will help you improve accessibility during the development cycle, such as Linters, Storybook, Playwright, just to name a few.
Focus: European Accessibility Act, Integrated Testing tools and methods (e.g. Linters, Storybook, Playwright)
Target audience: Everyone, Developers, Testers
Multistream in SIP and NoSIP @ OpenSIPS Summit 2025Lorenzo Miniero
Slides for my "Multistream support in the Janus SIP and NoSIP plugins" presentation at the OpenSIPS Summit 2025 event.
They describe my efforts refactoring the Janus SIP and NoSIP plugins to allow for the gatewaying of an arbitrary number of audio/video streams per call (thus breaking the current 1-audio/1-video limitation), plus some additional considerations on what this could mean when dealing with application protocols negotiated via SIP as well.
With Claude 4, Anthropic redefines AI capabilities, effectively unleashing a ...SOFTTECHHUB
With the introduction of Claude Opus 4 and Sonnet 4, Anthropic's newest generation of AI models is not just an incremental step but a pivotal moment, fundamentally reshaping what's possible in software development, complex problem-solving, and intelligent business automation.
Master tester AI toolbox - Kari Kakkonen at Testaus ja AI 2025 ProfessioKari Kakkonen
My slides at Professio Testaus ja AI 2025 seminar in Espoo, Finland.
Deck in English, even though I talked in Finnish this time, in addition to chairing the event.
I discuss the different motivations for testing to use AI tools to help in testing, and give several examples in each categories, some open source, some commercial.
Supercharge Your AI Development with Local LLMsFrancesco Corti
In today's AI development landscape, developers face significant challenges when building applications that leverage powerful large language models (LLMs) through SaaS platforms like ChatGPT, Gemini, and others. While these services offer impressive capabilities, they come with substantial costs that can quickly escalate especially during the development lifecycle. Additionally, the inherent latency of web-based APIs creates frustrating bottlenecks during the critical testing and iteration phases of development, slowing down innovation and frustrating developers.
This talk will introduce the transformative approach of integrating local LLMs directly into their development environments. By bringing these models closer to where the code lives, developers can dramatically accelerate development lifecycles while maintaining complete control over model selection and configuration. This methodology effectively reduces costs to zero by eliminating dependency on pay-per-use SaaS services, while opening new possibilities for comprehensive integration testing, rapid prototyping, and specialized use cases.
"AI in the browser: predicting user actions in real time with TensorflowJS", ...Fwdays
With AI becoming increasingly present in our everyday lives, the latest advancements in the field now make it easier than ever to integrate it into our software projects. In this session, we’ll explore how machine learning models can be embedded directly into front-end applications. We'll walk through practical examples, including running basic models such as linear regression and random forest classifiers, all within the browser environment.
Once we grasp the fundamentals of running ML models on the client side, we’ll dive into real-world use cases for web applications—ranging from real-time data classification and interpolation to object tracking in the browser. We'll also introduce a novel approach: dynamically optimizing web applications by predicting user behavior in real time using a machine learning model. This opens the door to smarter, more adaptive user experiences and can significantly improve both performance and engagement.
In addition to the technical insights, we’ll also touch on best practices, potential challenges, and the tools that make browser-based machine learning development more accessible. Whether you're a developer looking to experiment with ML or someone aiming to bring more intelligence into your web apps, this session will offer practical takeaways and inspiration for your next project.
cloudgenesis cloud workshop , gdg on campus mitasiyaldhande02
Step into the future of cloud computing with CloudGenesis, a power-packed workshop curated by GDG on Campus MITA, designed to equip students and aspiring cloud professionals with hands-on experience in Google Cloud Platform (GCP), Microsoft Azure, and Azure Al services.
This workshop offers a rare opportunity to explore real-world multi-cloud strategies, dive deep into cloud deployment practices, and harness the potential of Al-powered cloud solutions. Through guided labs and live demonstrations, participants will gain valuable exposure to both platforms- enabling them to think beyond silos and embrace a cross-cloud approach to
development and innovation.
Protecting Your Sensitive Data with Microsoft Purview - IRMS 2025Nikki Chapple
Session | Protecting Your Sensitive Data with Microsoft Purview: Practical Information Protection and DLP Strategies
Presenter | Nikki Chapple (MVP| Principal Cloud Architect CloudWay) & Ryan John Murphy (Microsoft)
Event | IRMS Conference 2025
Format | Birmingham UK
Date | 18-20 May 2025
In this closing keynote session from the IRMS Conference 2025, Nikki Chapple and Ryan John Murphy deliver a compelling and practical guide to data protection, compliance, and information governance using Microsoft Purview. As organizations generate over 2 billion pieces of content daily in Microsoft 365, the need for robust data classification, sensitivity labeling, and Data Loss Prevention (DLP) has never been more urgent.
This session addresses the growing challenge of managing unstructured data, with 73% of sensitive content remaining undiscovered and unclassified. Using a mountaineering metaphor, the speakers introduce the “Secure by Default” blueprint—a four-phase maturity model designed to help organizations scale their data security journey with confidence, clarity, and control.
🔐 Key Topics and Microsoft 365 Security Features Covered:
Microsoft Purview Information Protection and DLP
Sensitivity labels, auto-labeling, and adaptive protection
Data discovery, classification, and content labeling
DLP for both labeled and unlabeled content
SharePoint Advanced Management for workspace governance
Microsoft 365 compliance center best practices
Real-world case study: reducing 42 sensitivity labels to 4 parent labels
Empowering users through training, change management, and adoption strategies
🧭 The Secure by Default Path – Microsoft Purview Maturity Model:
Foundational – Apply default sensitivity labels at content creation; train users to manage exceptions; implement DLP for labeled content.
Managed – Focus on crown jewel data; use client-side auto-labeling; apply DLP to unlabeled content; enable adaptive protection.
Optimized – Auto-label historical content; simulate and test policies; use advanced classifiers to identify sensitive data at scale.
Strategic – Conduct operational reviews; identify new labeling scenarios; implement workspace governance using SharePoint Advanced Management.
🎒 Top Takeaways for Information Management Professionals:
Start secure. Stay protected. Expand with purpose.
Simplify your sensitivity label taxonomy for better adoption.
Train your users—they are your first line of defense.
Don’t wait for perfection—start small and iterate fast.
Align your data protection strategy with business goals and regulatory requirements.
💡 Who Should Watch This Presentation?
This session is ideal for compliance officers, IT administrators, records managers, data protection officers (DPOs), security architects, and Microsoft 365 governance leads. Whether you're in the public sector, financial services, healthcare, or education.
🔗 Read the blog: https://nikkichapple.com/irms-conference-2025/
SAP Sapphire 2025 ERP1612 Enhancing User Experience with SAP Fiori and AIPeter Spielvogel
Explore how AI in SAP Fiori apps enhances productivity and collaboration. Learn best practices for SAPUI5, Fiori elements, and tools to build enterprise-grade apps efficiently. Discover practical tips to deploy apps quickly, leveraging AI, and bring your questions for a deep dive into innovative solutions.
Introducing the OSA 3200 SP and OSA 3250 ePRCAdtran
Adtran's latest Oscilloquartz solutions make optical pumping cesium timing more accessible than ever. Discover how the new OSA 3200 SP and OSA 3250 ePRC deliver superior stability, simplified deployment and lower total cost of ownership. Built on a shared platform and engineered for scalable, future-ready networks, these models are ideal for telecom, defense, metrology and more.
GDG Cloud Southlake #43: Tommy Todd: The Quantum Apocalypse: A Looming Threat...James Anderson
The Quantum Apocalypse: A Looming Threat & The Need for Post-Quantum Encryption
We explore the imminent risks posed by quantum computing to modern encryption standards and the urgent need for post-quantum cryptography (PQC).
Bio: With 30 years in cybersecurity, including as a CISO, Tommy is a strategic leader driving security transformation, risk management, and program maturity. He has led high-performing teams, shaped industry policies, and advised organizations on complex cyber, compliance, and data protection challenges.
1. Tensor Flow
Tensors: n-dimensional arrays
A sequence of tensor operations
Deep learning process are flows of tensors
Vector: 1-D tensor
Matrix: 2-D tensor
Can represent also many machine learning algorithms
2. A simple ReLU network
a1 b1 c1
a0 b0 c0
w
a1=a0wa,a+b0wb,a+c0wc,a
b1=a0wa,b+b0wb,b+c0wc,b
c1=a0wa,c+b0wb,c+c0wc,c
Apply relu(…) on a1, b1, c1
Slower approach
Per-neuron operation
More efficient approach
Matrix operation
6. TensorFlow
Code so far defines a data flow graph
MatMul
ReLU
Variable
x
w = tf.Variable(tf.random_normal([3, 3]), name='w')
import tensorflow as tf
y = tf.matmul(x, w)
relu_out = tf.nn.relu(y)
Each variable corresponds to a
node in the graph, not the result
Can be confusing at the beginning
7. TensorFlow
Code so far defines a data flow graph
Needs to specify how we
want to execute the graph MatMul
ReLU
Variable
x
Session
Manage resource for graph execution
w = tf.Variable(tf.random_normal([3, 3]), name='w')
sess = tf.Session()
y = tf.matmul(x, w)
relu_out = tf.nn.relu(y)
import tensorflow as tf
result = sess.run(relu_out)
8. Graph
Fetch
Retrieve content from a node
w = tf.Variable(tf.random_normal([3, 3]), name='w')
sess = tf.Session()
y = tf.matmul(x, w)
relu_out = tf.nn.relu(y)
import tensorflow as tf
print sess.run(relu_out)
MatMul
ReLU
Variable
x
Fetch
We have assembled the pipes
Fetch the liquid
9. Graph
sess = tf.Session()
y = tf.matmul(x, w)
relu_out = tf.nn.relu(y)
import tensorflow as tf
print sess.run(relu_out)
sess.run(tf.initialize_all_variables())
w = tf.Variable(tf.random_normal([3, 3]), name='w')
InitializeVariable
Variable is an empty node
MatMul
ReLU
Variable
x
Fetch
Fill in the content of a
Variable node
10. Graph
sess = tf.Session()
y = tf.matmul(x, w)
relu_out = tf.nn.relu(y)
import tensorflow as tf
print sess.run(relu_out)
sess.run(tf.initialize_all_variables())
w = tf.Variable(tf.random_normal([3, 3]), name='w')
x = tf.placeholder("float", [1, 3])
Placeholder
How about x?
MatMul
ReLU
Variable
x
Fetch
placeholder(<data type>,
shape=<optional-shape>,
name=<optional-name>)
Its content will be fed
11. Graph
import numpy as np
import tensorflow as tf
sess = tf.Session()
x = tf.placeholder("float", [1, 3])
w = tf.Variable(tf.random_normal([3, 3]), name='w')
y = tf.matmul(x, w)
relu_out = tf.nn.relu(y)
sess.run(tf.initialize_all_variables())
print sess.run(relu_out, feed_dict={x:np.array([[1.0, 2.0, 3.0]])})
Feed
MatMul
ReLU
Variable
x
FetchPump liquid into the pipe
Feed
12. Session management
Needs to release resource after use
sess.close()
Common usage
with tf.Session() as sess:
…
Interactive
sess = InteractiveSession()
13. Prediction
import numpy as np
import tensorflow as tf
with tf.Session() as sess:
x = tf.placeholder("float", [1, 3])
w = tf.Variable(tf.random_normal([3, 3]), name='w')
relu_out = tf.nn.relu(tf.matmul(x, w))
softmax = tf.nn.softmax(relu_out)
sess.run(tf.initialize_all_variables())
print sess.run(softmax, feed_dict={x:np.array([[1.0, 2.0, 3.0]])})
Softmax
Make predictions for n targets that sum to 1
14. Prediction Difference
import numpy as np
import tensorflow as tf
with tf.Session() as sess:
x = tf.placeholder("float", [1, 3])
w = tf.Variable(tf.random_normal([3, 3]), name='w')
relu_out = tf.nn.relu(tf.matmul(x, w))
softmax = tf.nn.softmax(relu_out)
sess.run(tf.initialize_all_variables())
answer = np.array([[0.0, 1.0, 0.0]])
print answer - sess.run(softmax, feed_dict={x:np.array([[1.0, 2.0, 3.0]])})
15. Learn parameters: Loss
Define loss function
Loss function for softmax
softmax_cross_entropy_with_logits(
logits, labels, name=<optional-name>)
labels = tf.placeholder("float", [1, 3])
cross_entropy = tf.nn.softmax_cross_entropy_with_logits(
relu_out, labels, name='xentropy')
17. Iterative update
labels = tf.placeholder("float", [1, 3])
cross_entropy = tf.nn.softmax_cross_entropy_with_logits(
relu_out, labels, name=‘xentropy')
optimizer = tf.train.GradientDescentOptimizer(0.1)
train_op = optimizer.minimize(cross_entropy)
for step in range(10):
sess.run(train_op,
feed_dict= {x:np.array([[1.0, 2.0, 3.0]]), labels:answer})
Gradient descent usually needs more than one step
Run multiple times
18. Add parameters for Softmax
…
softmax_w = tf.Variable(tf.random_normal([3, 3]))
logit = tf.matmul(relu_out, softmax_w)
softmax = tf.nn.softmax(logit)
…
cross_entropy = tf.nn.softmax_cross_entropy_with_logits(
logit, labels, name=‘xentropy')
…
Do not want to use only non-negative input
Softmax layer
19. Add biases
…
w = tf.Variable(tf.random_normal([3, 3]))
b = tf.Variable(tf.zeros([1, 3]))
relu_out = tf.nn.relu(tf.matmul(x, w) + b)
softmax_w = tf.Variable(tf.random_normal([3, 3]))
softmax_b = tf.Variable(tf.zeros([1, 3]))
logit = tf.matmul(relu_out, softmax_w) + softmax_b
softmax = tf.nn.softmax(logit)
…
Biases initialized to zero
20. Make it deep
…
x = tf.placeholder("float", [1, 3])
relu_out = x
num_layers = 2
for layer in range(num_layers):
w = tf.Variable(tf.random_normal([3, 3]))
b = tf.Variable(tf.zeros([1, 3]))
relu_out = tf.nn.relu(tf.matmul(relu_out, w) + b)
…
Add layers
22. Improve naming, improve visualization
name_scope(name)
Help specify hierarchical names
…
for layer in range(num_layers):
with tf.name_scope('relu'):
w = tf.Variable(tf.random_normal([3, 3]))
b = tf.Variable(tf.zeros([1, 3]))
relu_out = tf.nn.relu(tf.matmul(relu_out, w) + b)
…
Will help visualizer to better
understand hierarchical relation
Move to outside the loop?
24. Add regularization to the loss
eg. L2 regularize on the Softmax layer parameters
…
l2reg = tf.reduce_sum(tf.square(softmax_w))
loss = cross_entropy + l2reg
train_op = optimizer.minimize(loss)
…
print sess.run(l2reg)
…
Add it to the loss
Automatic gradient calculation
29. Save and load models
tf.train.Saver(…)
Default will associate with all variables
all_variables()
save(sess, save_path, …)
restore(sess, save_path, …)
Replace initialization
That’s why we need to run initialization
separately
31. LSTM
# Parameters of gates are concatenated into one multiply for efficiency.
c, h = array_ops.split(1, 2, state)
concat = linear([inputs, h], 4 * self._num_units,True)
# i = input_gate, j = new_input, f = forget_gate, o = output_gate
i, j, f, o = array_ops.split(1, 4, concat)
new_c = c * sigmoid(f + self._forget_bias) + sigmoid(i) * tanh(j)
new_h = tanh(new_c) * sigmoid(o)
BasicLSTMCell
32. Word2Vec with TensorFlow
# Look up embeddings for inputs.
embeddings = tf.Variable(
tf.random_uniform([vocabulary_size, embedding_size], -1.0, 1.0))
embed = tf.nn.embedding_lookup(embeddings, train_inputs)
# Construct the variables for the NCE loss
nce_weights = tf.Variable(
tf.truncated_normal([vocabulary_size, embedding_size],
stddev=1.0 / math.sqrt(embedding_size)))
nce_biases = tf.Variable(tf.zeros([vocabulary_size]))
# Compute the average NCE loss for the batch.
# tf.nce_loss automatically draws a new sample of the negative labels each
# time we evaluate the loss.
loss = tf.reduce_mean(
tf.nn.nce_loss(nce_weights, nce_biases, embed, train_labels,
num_sampled, vocabulary_size))
33. Reuse Pre-trained models
Image recognition
Inception-v3
military uniform (866): 0.647296
suit (794): 0.0477196
academic gown (896): 0.0232411
bow tie (817): 0.0157356
bolo tie (940): 0.0145024
34. Try it on your Android
github.com/tensorflow/tensorflow/tree/master/tensorflow/
examples/android
Uses a Google Inception model to classify camera
frames in real-time, displaying the top results in an
overlay on the camera image.
Tensorflow Android Camera Demo
42. Google Brain Residency Program
Learn to conduct deep learning research w/experts in our team
Fixed one-year employment with salary, benefits, ...
Interesting problems,TensorFlow, and access to
computational resources
Goal after one year is to have conducted several research
projects
New one year immersion program in deep learning research
43. Google Brain Residency Program
Who should apply?
People with BSc, MSc or PhD, ideally in CS,
mathematics or statistics
Completed coursework in calculus, linear
algebra, and probability, or equiv.
Motivated, hard working, and have a strong
interest in deep learning
Programming experience
44. Google Brain Residency Program
Program Application & Timeline
DEADLINE: January 15, 2016
Thanks for your attention!