danielhanchen commited on
Commit
5260bcd
·
verified ·
1 Parent(s): 26be187

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,10 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ granite-3.3-2b-instruct-UD-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text
37
+ granite-3.3-2b-instruct-UD-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text
38
+ granite-3.3-2b-instruct-UD-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ granite-3.3-2b-instruct-UD-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text
40
+ granite-3.3-2b-instruct-UD-Q2_K_XL.gguf filter=lfs diff=lfs merge=lfs -text
41
+ granite-3.3-2b-instruct-UD-Q3_K_XL.gguf filter=lfs diff=lfs merge=lfs -text
42
+ granite-3.3-2b-instruct-UD-Q4_K_XL.gguf filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "GraniteForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attention_multiplier": 0.015625,
8
+ "bos_token_id": 0,
9
+ "embedding_multiplier": 12.0,
10
+ "eos_token_id": 0,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 2048,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 8192,
15
+ "logits_scaling": 8.0,
16
+ "max_position_embeddings": 131072,
17
+ "mlp_bias": false,
18
+ "model_type": "granite",
19
+ "num_attention_heads": 32,
20
+ "num_hidden_layers": 40,
21
+ "num_key_value_heads": 8,
22
+ "pad_token_id": 4,
23
+ "residual_multiplier": 0.22,
24
+ "rms_norm_eps": 1e-05,
25
+ "rope_scaling": null,
26
+ "rope_theta": 10000000.0,
27
+ "tie_word_embeddings": true,
28
+ "torch_dtype": "bfloat16",
29
+ "transformers_version": "4.52.0.dev0",
30
+ "unsloth_fixed": true,
31
+ "use_cache": true,
32
+ "vocab_size": 49159
33
+ }
granite-3.3-2b-instruct-UD-IQ1_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf8308500a8bda6c5c7b073ad9a197bf6c02821b4a0ec0f12723786e49f09c4c
3
+ size 657665408
granite-3.3-2b-instruct-UD-IQ1_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:125c3eda985a059bcd508d77e15dbb2d1f6f4591a1c43b810c53ede4549c96a3
3
+ size 612576640
granite-3.3-2b-instruct-UD-IQ2_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54784991716d2844aa3f3559cd30d2bc848e13b1381bd2ccedde3a675acc56ab
3
+ size 907570560
granite-3.3-2b-instruct-UD-IQ3_XXS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18a788dc917ac1cf44683ab60abf5a93404154072a9a12d61df073b42c88581d
3
+ size 1018457472
granite-3.3-2b-instruct-UD-Q2_K_XL.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11b42d4a2148e19d4589079d5b628b71a6a75715cf152aab33734b728d38bb20
3
+ size 1002910976
granite-3.3-2b-instruct-UD-Q3_K_XL.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c6bcc7276c16419f3796700567c458f1f1592a5c01fdfd60888a35438de9cca
3
+ size 1288033536
granite-3.3-2b-instruct-UD-Q4_K_XL.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab0adeef6265bede382a67f3a3108db6e9c27e72b021e1a12a5a5da4c2404217
3
+ size 1570731264