- ReLU activation function
- Stochastic Gradient Descent(SGD) optimizer
- Cross Entropy Loss Function
- Dataset: MNIST
- Validation data accuracy 97%
Validation Accuracy of Fully Connected NN
Validation Accuracy of CNN
- Model: ResNet50
- Optimizer: Adam
- Loss Function: Cross Entropy
Learning Rate is higher than expected but eventually, loss decreased enough
- Tensor Class: Included mathematical operations and ability to save gradients
- Implemented layer: Linear Layer
- Implemented Optimizers: SGD, Momentum, Adam, and RMSprop
- Implemented Loss Functions: Mean Squared Error, Categorical Cross Entropy
- Implemented Activation Functions: ReLU, LeakyReLU, Tanh, and Sigmoid