Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
## CIFAR-10 Results
|
2 |
-
In our
|
3 |
|
4 |
## Training Details
|
5 |
For each attention layer setting, we explored two model variants: an attention-only model and a model combining attention with the MLP module. Dropout and layer normalization were not applied for simplicity.
|
|
|
1 |
## CIFAR-10 Results
|
2 |
+
In our CIFAR-10 experiment, we aimed to assess the performance of Mice ViTs on a more complex and diverse dataset, CIFAR-10. We trained mice ViTs on classifying the 10 classes.
|
3 |
|
4 |
## Training Details
|
5 |
For each attention layer setting, we explored two model variants: an attention-only model and a model combining attention with the MLP module. Dropout and layer normalization were not applied for simplicity.
|