File size: 1,354 Bytes
a449177
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
---
library_name: PyTorch
tags:
- cnn
- lenet
- cifar100
- image-classification
datasets:
- uoft-cs/cifar100
language:
- en
metrics:
- accuracy
---

# CIFAR10 LeNet5 Variation 2: GELU + Dropout Layer

This repository contains our best variation of the original LeNet5 architecture adapted for CIFAR-10, but we will use its architecture and train it on CIFAR-100 this time. The model consists of two convolutional layers followed by two fully connected layers a dropout layer (p=0.5) and a final fully connected layer, using linear (GELU) activations, extending variation 1, and Kaiming uniform initialization. It is trained with a batch size of 32 using the Adam optimizer (learning rate 0.001) and CrossEntropyLoss. In our experiments, this model achieved a test loss of 0.0572 and a top-1 accuracy of 43.08% on CIFAR-100.

## Model Details

- **Architecture:** 2 Convolutional Layers, 2 Fully Connected Layers, 1 Dropout Layer, 1 Final Fully Connected Layer.
- **Activations:** GELU.
- **Weight Initialization:** Kaiming Uniform.
- **Optimizer:** Adam (lr=0.001).
- **Loss Function:** CrossEntropyLoss.
- **Dataset:** CIFAR-100.

## Usage

Load this model in PyTorch to fine-tune or evaluate on CIFAR-100 using your training and evaluation scripts.

---

Feel free to update this model card with further training details, benchmarks, or usage examples.