hubnemo HF Staff commited on
Commit
54fc349
·
verified ·
1 Parent(s): b07215f

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +1 -0
  2. adapter_config.json +4 -7
  3. adapter_model.safetensors +1 -1
README.md CHANGED
@@ -5,6 +5,7 @@ tags:
5
  - base_model:adapter:trl-internal-testing/tiny-Qwen2ForCausalLM-2.5
6
  - lora
7
  - transformers
 
8
  ---
9
 
10
  # Model Card for Model ID
 
5
  - base_model:adapter:trl-internal-testing/tiny-Qwen2ForCausalLM-2.5
6
  - lora
7
  - transformers
8
+ pipeline_tag: text-generation
9
  ---
10
 
11
  # Model Card for Model ID
adapter_config.json CHANGED
@@ -1,9 +1,6 @@
1
  {
2
  "alpha_pattern": {},
3
- "auto_mapping": {
4
- "base_model_class": "Qwen2ForCausalLM",
5
- "parent_library": "transformers.models.qwen2.modeling_qwen2"
6
- },
7
  "base_model_name_or_path": "trl-internal-testing/tiny-Qwen2ForCausalLM-2.5",
8
  "bias": "none",
9
  "corda_config": null,
@@ -27,10 +24,10 @@
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
30
- "v_proj",
31
- "q_proj"
32
  ],
33
- "task_type": null,
34
  "trainable_token_indices": null,
35
  "use_dora": false,
36
  "use_rslora": false
 
1
  {
2
  "alpha_pattern": {},
3
+ "auto_mapping": null,
 
 
 
4
  "base_model_name_or_path": "trl-internal-testing/tiny-Qwen2ForCausalLM-2.5",
5
  "bias": "none",
6
  "corda_config": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "q_proj",
28
+ "v_proj"
29
  ],
30
+ "task_type": "CAUSAL_LM",
31
  "trainable_token_indices": null,
32
  "use_dora": false,
33
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a266dab84463d41ee776973d1808bb75d1050d0a615922c6402781fa20e09206
3
  size 2800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a790a1eb431da10048893ce1c1c8947b6495695ff635a1653f5f6c36beee8cbb
3
  size 2800