Deep Learning Blooms Taxonomy Questions Deep Learning Assignment
UNIT I: Machine Learning Basics & Deep Feedforward Networks
Machine Learning Fundamentals
- Learning Algorithms – Types and Frameworks
- Model Capacity – Underfitting vs. Overfitting
- Hyperparameters and Validation Sets – Train/Validation/Test Split, Cross-Validation
- Estimators – Point Estimation, Interval Estimation
- Bias-Variance Tradeoff – Decomposition of Generalization Error
- Maximum Likelihood Estimation (MLE) – Principles and Applications
- Bayesian Statistics – Priors, Posteriors, MAP Estimation
- Supervised Learning Algorithms – Regression, Classification (k-NN, SVM, Decision Trees)
- Unsupervised Learning Algorithms – Clustering (k-Means), Dimensionality Reduction (PCA)
- Stochastic Gradient Descent (SGD) – Mini-batch, Convergence, Momentum
- Building a Machine Learning Algorithm – Pipeline: Data → Features → Model → Evaluation
- Challenges Motivating Deep Learning – Curse of Dimensionality, Feature Engineering Bottlenecks
Deep Feedforward Networks (Multilayer Perceptrons)
- Learning XOR – Limitations of Linear Models, Need for Hidden Layers
- Gradient-Based Learning – Chain Rule, Computational Graphs
- Hidden Units – Activation Functions: ReLU, Sigmoid, Tanh, Softmax
- Architecture Design – Depth vs. Width, Universal Approximation Theorem
- Backpropagation and Other Differentiation Algorithms
- Automatic Differentiation
- Forward vs. Reverse Mode
- Jacobian and Hessian Matrices
UNIT II: Regularization & Optimization for Deep Learning
Regularization Techniques
- Parameter Norm Penalties – L1 (Lasso), L2 (Ridge), Elastic Net
- Norm Penalties as Constrained Optimization – Lagrange Multipliers Interpretation
- Regularization and Under-Constrained Problems – Ill-posed Problems, Tikhonov Regularization
- Dataset Augmentation – Image Flips, Rotations, Noise Injection
- Noise Robustness – Input/Weight Noise, Denoising Autoencoders
- Semi-Supervised Learning – Leveraging Unlabeled Data
- Multi-Task Learning – Shared Representations, Auxiliary Tasks
- Early Stopping – Validation-Based Halting Criterion
- Parameter Tying and Parameter Sharing – CNNs, RNNs, Weight Reuse
- Sparse Representations – L1 Regularization, Sparse Coding
- Bagging and Other Ensemble Methods – Random Forests, Boosting (AdaBoost, XGBoost)
- Dropout – Training-Time Neuron Deactivation, Inference Scaling
- Adversarial Training – Robustness to Perturbations, FGSM
- Tangent Distance, Tangent Prop, and Manifold Tangent Classifier – Invariance Learning
Optimization for Training Deep Models
- Learning vs Pure Optimization – Generalization ≠ Minimizing Training Loss
- Challenges in Neural Network Optimization
- Poor Conditioning, Local Minima, Saddle Points, Vanishing/Exploding Gradients
- Basic Algorithms – SGD, Momentum, Nesterov Momentum
- Parameter Initialization Strategies
- Xavier/Glorot, He Initialization, Orthogonal Initialization
- Algorithms with Adaptive Learning Rates
- AdaGrad, RMSProp, Adam, Nadam
UNIT III: Convolutional Networks
Fundamentals of CNNs
- The Convolution Operation – Kernels, Feature Maps, Stride, Padding
- Motivation – Translation Invariance, Parameter Sharing, Hierarchical Features
- Pooling – Max Pooling, Average Pooling, Global Pooling, Invariance Effects
- Convolution and Pooling as an Infinitely Strong Prior – Spatial Locality, Stationarity
- Variants of the Basic Convolution Function
- Dilated (Atrous) Convolutions
- Depthwise Separable Convolutions
- Transposed Convolutions (Deconvolutions)
Advanced Topics
- Structured Outputs – Semantic Segmentation, Object Detection Architectures
- Data Types – Images, Volumes (3D CNNs), Sequences (1D CNNs)
- Efficient Convolution Algorithms – FFT-based, Winograd, Im2Col
- Random or Unsupervised Features – Random CNNs, Self-Supervised Pretext Tasks
UNIT IV: Recurrent and Recursive Networks
Sequence Modeling
- Unfolding Computational Graphs – Time-Unrolled Representations
- Recurrent Neural Networks (RNNs) – Vanilla RNN, Hidden State Dynamics
- Bidirectional RNNs – Forward + Backward Context Integration
- Encoder-Decoder Sequence-to-Sequence Architectures – Machine Translation, Text Summarization
- Deep Recurrent Networks – Stacked RNNs, Residual Connections
- Recursive Neural Networks – Tree-Structured Inputs (e.g., Parse Trees)
Long-Term Dependency Challenges
- The Challenge of Long-Term Dependencies – Gradient Vanishing in Vanilla RNNs
- Echo State Networks (ESNs) – Fixed Recurrent Weights, Readout Training
- Leaky Units and Other Strategies for Multiple Time Scales – Time Constants, Hierarchical RNNs
- The Long Short-Term Memory (LSTM) and Other Gated RNNs
- LSTM: Input, Forget, Output Gates
- GRU (Gated Recurrent Unit): Simplified Gating
- Optimization for Long-Term Dependencies – Gradient Clipping, Curriculum Learning
- Explicit Memory – Neural Turing Machines, Memory Networks, Differentiable Neural Computers
UNIT V: Practical Methodology & Applications
Practical Methodology
- Performance Metrics
- Classification: Accuracy, Precision, Recall, F1, AUC-ROC
- Regression: MSE, MAE, R²
- Ranking: NDCG, MAP
- Default Baseline Models – Logistic Regression, Random Forest, Linear SVM
- Determining Whether to Gather More Data – Learning Curves, Data Saturation
- Selecting Hyperparameters – Grid Search, Random Search, Bayesian Optimization
- Debugging Strategies – Gradient Checks, Activation/Gradient Distributions, Overfitting Diagnosis
- Example: Multi-Digit Number Recognition – End-to-End System Design, Error Analysis
Applications of Deep Learning
- Large-Scale Deep Learning – Distributed Training, Model Parallelism, Data Parallelism
- Computer Vision
- Image Classification (ResNet, EfficientNet)
- Object Detection (YOLO, Faster R-CNN)
- Segmentation (U-Net, Mask R-CNN)
- Speech Recognition
- End-to-End Models (DeepSpeech, Wav2Vec)
- CTC Loss, Attention Models
- Natural Language Processing (NLP)
- Transformers (BERT, GPT)
- Named Entity Recognition, Machine Translation, Question Answering
- Other Applications
- Recommender Systems
- Generative Models (GANs, VAEs, Diffusion Models)
- Reinforcement Learning (Deep Q-Networks, Policy Gradients)
- Healthcare, Finance, Robotics, Autonomous Systems