RGPV Notes _ Machine Learning
RGPV Notes _ Machine Learning
Linearity vs non-linearity
Activation functions (e.g., sigmoid, ReLU)
Weights and bias
Loss function
Gradient descent
Multilayer networks
Backpropagation
Weight initialization
Flattening
Subsampling
Padding
Stride
Convolutional layer
Pooling layer
Loss layer
1×1 convolution
Inception network
Input channels
Transfer learning
One-shot learning
Dimension reduction
Implementation using frameworks like TensorFlow, Keras, etc.
Related posts: