0% found this document useful (0 votes)
53 views15 pages

A145286344 23681 24 2018 Tensorflow

The document discusses key topics in intelligent systems including soft computing techniques, machine learning foundations and advanced machine learning. It then summarizes capabilities and applications of SciPy, Scikit-learn, Tensorflow and Keras including clustering algorithms, interpolation, image processing, deep learning models and more.

Uploaded by

Rohit Kolli
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
53 views15 pages

A145286344 23681 24 2018 Tensorflow

The document discusses key topics in intelligent systems including soft computing techniques, machine learning foundations and advanced machine learning. It then summarizes capabilities and applications of SciPy, Scikit-learn, Tensorflow and Keras including clustering algorithms, interpolation, image processing, deep learning models and more.

Uploaded by

Rohit Kolli
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
You are on page 1/ 15

Intelligent Systems Track

• Soft Computing Techniques


Neural Networks, Fuzzy logic, Genetic Algorithm
• Machine Learning Foundation
Supervised , Unsupervised, Reinforcement learning
• Advanced Machine Learning
Tensorflow, keras etc..
• Pattern Recognition
scikit
• SciPy, pronounced as Sigh Pi, is a scientific
python open source, distributed under the
BSD licensed library to perform Mathematical,
Scientific and Engineering Computations.
• The SciPy library depends on NumPy, which
provides convenient and fast N-dimensional
array manipulation.
SciPy - Cluster
• K-means clustering is a method for finding clusters and cluster
centers in a set of unlabeled data.
• Intuitively, we might think of a cluster as – comprising of a group of
data points, whose inter-point distances are small compared with
the distances to points outside of the cluster.
• Given an initial set of K centers, the K-means algorithm iterates the
following two steps −
– For each center, the subset of training points (its cluster) that is closer
to it is identified than any other center.
– The mean of each feature for the data points in each cluster are
computed, and this mean vector becomes the new center for that
cluster.
• These two steps are iterated until the centers no longer move or
the assignments no longer change.
• Then, a new point x can be assigned to the cluster of the closest
prototype.
SciPy - Interpolate
• Interpolation is the process of finding a value between two points
on a line or a curve.
• To help us remember what it means, we should think of the first
part of the word, 'inter,' as meaning 'enter,' which reminds us to
look 'inside' the data we originally had.
Program
import numpy as np
from scipy import interpolate
import matplotlib.pyplot as plt
x = np.linspace(0, 4, 12)
y = np.cos(x**2/3+4)
print (x,y)
plt.plot(x, y,’o’)
plt.show()
SciPy - Ndimage
• The SciPy ndimage submodule is dedicated to image
processing. Here, ndimage means an n-dimensional image.
• Some of the most common tasks in image processing are as
follows &miuns;
– Input/Output, displaying images
– Basic manipulations − Cropping, flipping, rotating, etc.
– Image filtering − De-noising, sharpening, etc.
– Image segmentation − Labeling pixels corresponding to different
objects
– Classification
– Feature extraction
– Registration
Neural Network
Tensorflow
• TensorFlow is a library based on Python that
provides different types of functionality for
implementing Deep Learning Models.
• In TensorFlow, the term tensor refers to the
representation of data as multi-dimensional
array whereas the term flow refers to the
series of operations that one performs on
tensors
Tensorflow
• The overall process of writing a TensorFlow
program involves two steps:
Building a Computational Graph
Running a Computational Graph
Building a Computational Graph
Running a Computational Graph
• A session encapsulates the control and state
of the TensorFlow runtime i.e. it stores the
information about the order in which all the
operations will be performed and passes the
result of already computed operation to the
next operation in the pipeline.
• Let me show you how to run the above
computational graph within a session
(Explanation of each line of code has been
added as a comment):
Running a Computational Graph
Application of Tensorflow
• It’s a powerful machine learning framework
• It is used for deep learning.
• It helps classify and cluster data like that with sometimes
superhuman accuracy.
• RankBrain: A large-scale deployment of deep neural nets for search
ranking on Google
• SmartReply: Deep LSTM model to automatically generate email
responses
• On-Device Computer Vision for OCR: On-device computer vision
model to do optical character recognition to enable real-time
translation.
• Massively Multitask Networks for Drug Discovery: A deep neural
network model for identifying promising drug candidates.
Application of Keras
• Keras Applications are deep learning models that are made available alongside
pre-trained weights. These models can be used for prediction, feature extraction,
and fine-tuning.
• Weights are downloaded automatically when instantiating a model.
• Models for image classification with weights trained on ImageNet:
– Xception
– VGG16
– VGG19
– ResNet50
– InceptionV3
– InceptionResNetV2
– MobileNet
– DenseNet
– NASNet
– MobileNetV2

• Keras has broad adoption in the industry and the research community
Application of Scikit-Learn
• Simple and efficient tools for data mining and data analysis
• Built on NumPy, SciPy, and matplotlib
• Applications:
– Classification: identifying to which category an object belongs
to.
– Regression: Predicting a continuous-valued attribute associated
with an object.
– Clustering: Automatic grouping of similar objects into sets.
– Dimensionality reduction: Reducing the number of random
variables to consider.
– Model Selection: Comparing, validating and choosing
parameters and models.
– Preprocessing: Feature extraction and normalization.

You might also like