rwood-97 commited on
Commit
79b77a7
·
verified ·
1 Parent(s): d4e9ce4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -2
README.md CHANGED
@@ -18,7 +18,7 @@ datasets:
18
  - Livingwithmachines/MapReader_Data_SIGSPATIAL_2022
19
  ---
20
 
21
- # Model card for mr_vit_base_patch16_224_timm_pretrain
22
 
23
  A Vision Transformer (ViT) model pre-trained on ImageNet-21k (14 million images, 21,843 classes) at resolution 224x224, and fine-tuned on ImageNet 2012 (1 million images, 1,000 classes) at resolution 224x224.
24
  Fine-tuned on gold standard annotations and outputs from early experiments using MapReader (found [here](https://huggingface.co/datasets/Livingwithmachines/MapReader_Data_SIGSPATIAL_2022)).
@@ -30,9 +30,19 @@ Fine-tuned on gold standard annotations and outputs from early experiments using
30
  - **Model type:** Image classification /feature backbone
31
  - **Finetuned from model:** https://huggingface.co/google/vit-base-patch16-224
32
 
 
 
 
 
 
 
 
33
  ## Uses
34
 
35
- This fine-tuned version of the model is an output of the MapReader pipeline. It was used to classify 'patch' images (cells/regions) of scanned nineteenth-century series maps of Britain provided by the National Library of Scotland (learn more [here](https://maps.nls.uk/os/)). We classified patches to indicate the presence of buildings and railway infrastructure. See [our paper](https://dl.acm.org/doi/10.1145/3557919.3565812) for more details about labels.
 
 
 
36
 
37
  ## How to Get Started with the Model in MapReader
38
 
 
18
  - Livingwithmachines/MapReader_Data_SIGSPATIAL_2022
19
  ---
20
 
21
+ # Model card for mr_vit_base_patch16_224_timm_pretrain_railspace_and_building
22
 
23
  A Vision Transformer (ViT) model pre-trained on ImageNet-21k (14 million images, 21,843 classes) at resolution 224x224, and fine-tuned on ImageNet 2012 (1 million images, 1,000 classes) at resolution 224x224.
24
  Fine-tuned on gold standard annotations and outputs from early experiments using MapReader (found [here](https://huggingface.co/datasets/Livingwithmachines/MapReader_Data_SIGSPATIAL_2022)).
 
30
  - **Model type:** Image classification /feature backbone
31
  - **Finetuned from model:** https://huggingface.co/google/vit-base-patch16-224
32
 
33
+ ### Classes and labels
34
+
35
+ - 0: no
36
+ - 1: railspace
37
+ - 2: building
38
+ - 3: railspace & building
39
+
40
  ## Uses
41
 
42
+ This fine-tuned version of the model is an output of the MapReader pipeline.
43
+ It was used to classify 'patch' images (cells/regions) of scanned nineteenth-century series maps of Britain provided by the National Library of Scotland (learn more [here](https://maps.nls.uk/os/)).
44
+ We classified patches to indicate the presence of buildings and railway infrastructure.
45
+ See [our paper](https://dl.acm.org/doi/10.1145/3557919.3565812) for more details about labels.
46
 
47
  ## How to Get Started with the Model in MapReader
48