Files changed (1) hide show
  1. allenai_MolmoE-1B-0924.json +127 -0
allenai_MolmoE-1B-0924.json ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bomFormat": "CycloneDX",
3
+ "specVersion": "1.6",
4
+ "serialNumber": "urn:uuid:b374c80a-3be7-4159-ab78-2d78ffffdeba",
5
+ "version": 1,
6
+ "metadata": {
7
+ "timestamp": "2025-06-05T09:36:27.616407+00:00",
8
+ "component": {
9
+ "type": "machine-learning-model",
10
+ "bom-ref": "allenai/MolmoE-1B-0924-fa3b3569-ea6e-5505-8855-0af97b179252",
11
+ "name": "allenai/MolmoE-1B-0924",
12
+ "externalReferences": [
13
+ {
14
+ "url": "https://huggingface.co/allenai/MolmoE-1B-0924",
15
+ "type": "documentation"
16
+ }
17
+ ],
18
+ "modelCard": {
19
+ "modelParameters": {
20
+ "task": "image-text-to-text",
21
+ "architectureFamily": "molmo",
22
+ "modelArchitecture": "OLMoForCausalLM",
23
+ "datasets": [
24
+ {
25
+ "ref": "allenai/OLMoE-mix-0924-9134d812-8ebc-5945-9ce1-043f54ccbce5"
26
+ }
27
+ ]
28
+ },
29
+ "properties": [
30
+ {
31
+ "name": "library_name",
32
+ "value": "transformers"
33
+ },
34
+ {
35
+ "name": "base_model",
36
+ "value": "openai/clip-vit-large-patch14-336, allenai/OLMoE-1B-7B-0924"
37
+ }
38
+ ]
39
+ },
40
+ "authors": [
41
+ {
42
+ "name": "allenai"
43
+ }
44
+ ],
45
+ "licenses": [
46
+ {
47
+ "license": {
48
+ "id": "Apache-2.0",
49
+ "url": "https://spdx.org/licenses/Apache-2.0.html"
50
+ }
51
+ }
52
+ ],
53
+ "tags": [
54
+ "transformers",
55
+ "pytorch",
56
+ "molmo",
57
+ "text-generation",
58
+ "multimodal",
59
+ "moe",
60
+ "olmo",
61
+ "olmoe",
62
+ "molmoe",
63
+ "image-text-to-text",
64
+ "custom_code",
65
+ "en",
66
+ "dataset:allenai/OLMoE-mix-0924",
67
+ "arxiv:2409.17146",
68
+ "base_model:allenai/OLMoE-1B-7B-0924",
69
+ "base_model:finetune:allenai/OLMoE-1B-7B-0924",
70
+ "license:apache-2.0",
71
+ "autotrain_compatible",
72
+ "endpoints_compatible",
73
+ "region:us"
74
+ ]
75
+ }
76
+ },
77
+ "components": [
78
+ {
79
+ "type": "data",
80
+ "bom-ref": "allenai/OLMoE-mix-0924-9134d812-8ebc-5945-9ce1-043f54ccbce5",
81
+ "name": "allenai/OLMoE-mix-0924",
82
+ "data": [
83
+ {
84
+ "type": "dataset",
85
+ "bom-ref": "allenai/OLMoE-mix-0924-9134d812-8ebc-5945-9ce1-043f54ccbce5",
86
+ "name": "allenai/OLMoE-mix-0924",
87
+ "contents": {
88
+ "url": "https://huggingface.co/datasets/allenai/OLMoE-mix-0924",
89
+ "properties": [
90
+ {
91
+ "name": "task_categories",
92
+ "value": "text-generation"
93
+ },
94
+ {
95
+ "name": "language",
96
+ "value": "en"
97
+ },
98
+ {
99
+ "name": "size_categories",
100
+ "value": "1B<n<10B"
101
+ },
102
+ {
103
+ "name": "pretty_name",
104
+ "value": "OLMoE Mix (September 2024)"
105
+ },
106
+ {
107
+ "name": "license",
108
+ "value": "odc-by"
109
+ }
110
+ ]
111
+ },
112
+ "governance": {
113
+ "owners": [
114
+ {
115
+ "organization": {
116
+ "name": "allenai",
117
+ "url": "https://huggingface.co/allenai"
118
+ }
119
+ }
120
+ ]
121
+ },
122
+ "description": "\n\t\n\t\t\n\t\tOLMoE Mix (September 2024)\n\t\n\n\n\nThe following data mix was used to train OLMoE-1B-7B, a Mixture-of-Experts LLM with 1B active and 7B total parameters released in September 2024. \nThe base version of OLMoE-1B-7B can be found at this page, the SFT of OLMoE-1B-7B is available here, and a version combining SFT and DPO is available following this link.\n\n\t\n\t\t\n\t\n\t\n\t\tStatistics\n\t\n\n\n\t\n\t\t\nSubset\nTokens\nWords\nBytes\nDocs\n\n\n\t\t\nDCLM Baseline 1.0\n3.86 T\n3.38 T16.7 T\n2.95 B\n\n\nStarcoder\n101 B\n63.9 B\u2026 See the full description on the dataset page: https://huggingface.co/datasets/allenai/OLMoE-mix-0924."
123
+ }
124
+ ]
125
+ }
126
+ ]
127
+ }