Triangle104 commited on
Commit
2eeebc1
·
verified ·
1 Parent(s): a9688f7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -12
README.md CHANGED
@@ -1,20 +1,51 @@
1
- ---
2
- base_model: THUDM/GLM-Z1-Rumination-32B-0414
3
- language:
4
- - zh
5
- - en
6
- library_name: transformers
7
- license: mit
8
- pipeline_tag: text-generation
9
- tags:
10
- - llama-cpp
11
- - gguf-my-repo
12
- ---
13
 
14
  # Triangle104/GLM-Z1-Rumination-32B-0414-Q6_K-GGUF
15
  This model was converted to GGUF format from [`THUDM/GLM-Z1-Rumination-32B-0414`](https://huggingface.co/THUDM/GLM-Z1-Rumination-32B-0414) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/THUDM/GLM-Z1-Rumination-32B-0414) for more details on the model.
17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ## Use with llama.cpp
19
  Install llama.cpp through brew (works on Mac and Linux)
20
 
 
1
+ ---
2
+ base_model: THUDM/GLM-Z1-Rumination-32B-0414
3
+ language:
4
+ - zh
5
+ - en
6
+ library_name: transformers
7
+ license: mit
8
+ pipeline_tag: text-generation
9
+ tags:
10
+ - llama-cpp
11
+ - gguf-my-repo
12
+ ---
13
 
14
  # Triangle104/GLM-Z1-Rumination-32B-0414-Q6_K-GGUF
15
  This model was converted to GGUF format from [`THUDM/GLM-Z1-Rumination-32B-0414`](https://huggingface.co/THUDM/GLM-Z1-Rumination-32B-0414) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/THUDM/GLM-Z1-Rumination-32B-0414) for more details on the model.
17
 
18
+ ---
19
+ Introduction
20
+ -
21
+ The GLM family welcomes a new generation of open-source models, the GLM-4-32B-0414
22
+ series, featuring 32 billion parameters. Its performance is comparable
23
+ to OpenAI's GPT series and DeepSeek's V3/R1 series, and it supports very
24
+ user-friendly local deployment features. GLM-4-32B-Base-0414 was
25
+ pre-trained on 15T of high-quality data, including a large amount of
26
+ reasoning-type synthetic data, laying the foundation for subsequent
27
+ reinforcement learning extensions. In the post-training stage, in
28
+ addition to human preference alignment for dialogue scenarios, we also
29
+ enhanced the model's performance in instruction following, engineering
30
+ code, and function calling using techniques such as rejection sampling
31
+ and reinforcement learning, strengthening the atomic capabilities
32
+ required for agent tasks. GLM-4-32B-0414 achieves good results in areas
33
+ such as engineering code, Artifact generation, function calling,
34
+ search-based Q&A, and report generation. Some benchmarks even rival
35
+ larger models like GPT-4o and DeepSeek-V3-0324 (671B).
36
+
37
+ GLM-Z1-Rumination-32B-0414 is a deep reasoning model with rumination capabilities
38
+ (benchmarked against OpenAI's Deep Research). Unlike typical deep
39
+ thinking models, the rumination model employs longer periods of deep
40
+ thought to solve more open-ended and complex problems (e.g., writing a
41
+ comparative analysis of AI development in two cities and their future
42
+ development plans). The rumination model integrates search tools during
43
+ its deep thinking process to handle complex tasks and is trained by
44
+ utilizing multiple rule-based rewards to guide and extend end-to-end
45
+ reinforcement learning. Z1-Rumination shows significant improvements in
46
+ research-style writing and complex retrieval tasks.
47
+
48
+ ---
49
  ## Use with llama.cpp
50
  Install llama.cpp through brew (works on Mac and Linux)
51