Syllabus
Overview of AI, ML, and DL, Multi-layer perceptrons, Activation functions (ReLU, Sigmoid, Tanh), Loss functions and cost, Performance analysis of Classifier and Regression model. Essential Data Pre-processing for Deep Learning.
Backpropagation and gradient descent, Stochastic Gradient Descent (SGD), RMSProp, Adam. Weight initialization. Overfitting and underfitting. Regularization techniques: regularization, dropout, Batch normalization. Deep Neural Networks.
Convolution layers, pooling layers, CNN architectures. 1D CNNs for signal processing, Transfer learning and fine-tuning pre-trained models, Visualization techniques: saliency maps, Grad-CAM for explaining CNN decisions, Application of CNN.
Comparison between static and sequential data, Sequence modeling, RNN architecture, Limitations of RNNs: vanishing/exploding gradients, short memory retention, Internal structure of LSTM, Gated Recurrent Unit (GRU) Networks, Comparison in terms of performance and complexity, Advanced architectures: GAN, Transformers.