Xenova HF Staff commited on
Commit
bc2ee71
·
verified ·
1 Parent(s): 2e8fdb5

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_attn_implementation_autoset": true,
3
- "_name_or_path": "gg-hf-g/gemma-3-1b-it",
4
  "architectures": [
5
  "Gemma3ForCausalLM"
6
  ],
@@ -9,7 +9,10 @@
9
  "attn_logit_softcapping": null,
10
  "bos_token_id": 2,
11
  "cache_implementation": "hybrid",
12
- "eos_token_id": 1,
 
 
 
13
  "final_logit_softcapping": null,
14
  "head_dim": 256,
15
  "hidden_activation": "gelu_pytorch_tanh",
@@ -31,11 +34,6 @@
31
  "sliding_window_pattern": 6,
32
  "torch_dtype": "float32",
33
  "transformers_version": "4.50.0.dev0",
34
- "transformers.js_config": {
35
- "use_external_data_format": {
36
- "model.onnx": true
37
- }
38
- },
39
  "use_cache": true,
40
  "vocab_size": 262144
41
  }
 
1
  {
2
  "_attn_implementation_autoset": true,
3
+ "_name_or_path": "google/gemma-3-1b-it",
4
  "architectures": [
5
  "Gemma3ForCausalLM"
6
  ],
 
9
  "attn_logit_softcapping": null,
10
  "bos_token_id": 2,
11
  "cache_implementation": "hybrid",
12
+ "eos_token_id": [
13
+ 1,
14
+ 106
15
+ ],
16
  "final_logit_softcapping": null,
17
  "head_dim": 256,
18
  "hidden_activation": "gelu_pytorch_tanh",
 
34
  "sliding_window_pattern": 6,
35
  "torch_dtype": "float32",
36
  "transformers_version": "4.50.0.dev0",
 
 
 
 
 
37
  "use_cache": true,
38
  "vocab_size": 262144
39
  }
generation_config.json CHANGED
@@ -2,7 +2,10 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 2,
4
  "cache_implementation": "hybrid",
5
- "eos_token_id": [1, 106],
 
 
 
6
  "pad_token_id": 0,
7
  "transformers_version": "4.50.0.dev0"
8
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 2,
4
  "cache_implementation": "hybrid",
5
+ "eos_token_id": [
6
+ 1,
7
+ 106
8
+ ],
9
  "pad_token_id": 0,
10
  "transformers_version": "4.50.0.dev0"
11
  }
onnx/model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19fe4f234ac17c6a6fbd3e322684bb3ee4165f440853844ec9411fa4730c4c00
3
- size 1011486
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fcdd6cc5892fb4f3ea04b423657561b437239b37ab963b85bebae4532081a16f
3
+ size 1012725
tokenizer_config.json CHANGED
@@ -51337,7 +51337,6 @@
51337
  "image_token": "<image_soft_token>",
51338
  "model_max_length": 1000000000000000019884624838656,
51339
  "pad_token": "<pad>",
51340
- "processor_class": "Gemma3Processor",
51341
  "sp_model_kwargs": null,
51342
  "spaces_between_special_tokens": false,
51343
  "tokenizer_class": "GemmaTokenizer",
 
51337
  "image_token": "<image_soft_token>",
51338
  "model_max_length": 1000000000000000019884624838656,
51339
  "pad_token": "<pad>",
 
51340
  "sp_model_kwargs": null,
51341
  "spaces_between_special_tokens": false,
51342
  "tokenizer_class": "GemmaTokenizer",