LG-AI-EXAONE commited on
Commit
48023ad
·
1 Parent(s): 6da3418

Update HuggingFace transformers support

Browse files
Files changed (3) hide show
  1. README.md +2 -8
  2. config.json +33 -1
  3. generation_config.json +1 -1
README.md CHANGED
@@ -20,6 +20,7 @@ library_name: transformers
20
  <img src="assets/EXAONE_Symbol+BI_3d.png", width="300", style="margin: 40 auto;">
21
  🎉 License Updated! We are pleased to announce our more flexible licensing terms 🤗
22
  <br>✈️ Try on <a href="https://friendli.ai/suite/~/serverless-endpoints/LGAI-EXAONE/EXAONE-4.0-32B/overview">FriendliAI</a>
 
23
  <br>
24
 
25
  # EXAONE-4.0-1.2B-GPTQ-Int8
@@ -50,14 +51,7 @@ For more details, please refer to our [technical report](https://arxiv.org/abs/2
50
 
51
  ## Quickstart
52
 
53
- You should install the transformers library forked from the original, available in our [PR](https://github.com/huggingface/transformers/pull/39129).
54
- Once this PR is merged and released, we will update this section.
55
-
56
- You can install the latest version of transformers with support for EXAONE 4.0 by following the command:
57
-
58
- ```bash
59
- pip install git+https://github.com/lgai-exaone/transformers@add-exaone4
60
- ```
61
 
62
  ### Non-reasoning mode
63
 
 
20
  <img src="assets/EXAONE_Symbol+BI_3d.png", width="300", style="margin: 40 auto;">
21
  🎉 License Updated! We are pleased to announce our more flexible licensing terms 🤗
22
  <br>✈️ Try on <a href="https://friendli.ai/suite/~/serverless-endpoints/LGAI-EXAONE/EXAONE-4.0-32B/overview">FriendliAI</a>
23
+ <br><br><i>📢 EXAONE 4.0 is officially supported by HuggingFace transformers! Please check out the guide <a href="#quickstart">below</a></i>
24
  <br>
25
 
26
  # EXAONE-4.0-1.2B-GPTQ-Int8
 
51
 
52
  ## Quickstart
53
 
54
+ You should install the transformers library with version >= `4.54.0`.
 
 
 
 
 
 
 
55
 
56
  ### Non-reasoning mode
57
 
config.json CHANGED
@@ -10,6 +10,38 @@
10
  "hidden_size": 2048,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 4096,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  "max_position_embeddings": 65536,
14
  "model_type": "exaone4",
15
  "num_attention_heads": 32,
@@ -50,7 +82,7 @@
50
  "sliding_window_pattern": null,
51
  "tie_word_embeddings": true,
52
  "torch_dtype": "bfloat16",
53
- "transformers_version": "4.54.0.dev0",
54
  "use_cache": false,
55
  "vocab_size": 102400
56
  }
 
10
  "hidden_size": 2048,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 4096,
13
+ "layer_types": [
14
+ "full_attention",
15
+ "full_attention",
16
+ "full_attention",
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention"
44
+ ],
45
  "max_position_embeddings": 65536,
46
  "model_type": "exaone4",
47
  "num_attention_heads": 32,
 
82
  "sliding_window_pattern": null,
83
  "tie_word_embeddings": true,
84
  "torch_dtype": "bfloat16",
85
+ "transformers_version": "4.54.0",
86
  "use_cache": false,
87
  "vocab_size": 102400
88
  }
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 361,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.54.0.dev0"
7
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 361,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.54.0"
7
  }