ZeroXClem commited on
Commit
636580b
·
verified ·
1 Parent(s): 8d18c0f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +107 -6
README.md CHANGED
@@ -4,17 +4,51 @@ tags:
4
  - merge
5
  - mergekit
6
  - lazymergekit
 
 
 
 
 
 
 
 
 
 
7
  ---
8
 
9
- # ZeroXClem-Llama-3.1-8B-Athena-Apollo-exp
10
 
11
- ZeroXClem-Llama-3.1-8B-Athena-Apollo-exp is a merge of the following models using [mergekit](https://github.com/cg123/mergekit):
12
 
13
- ## 🧩 Configuration
14
 
15
- ```yaml
16
- # Merge configuration for ZeroXClem-Llama-3.1-8B-Athena-Apollo-exp using MODEL STOCK
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
 
 
 
 
18
  name: ZeroXClem-Llama-3.1-8B-Athena-Apollo-exp
19
  base_model: mergekit-community/L3.1-Athena-l3-8B
20
  dtype: bfloat16
@@ -25,4 +59,71 @@ models:
25
  - model: mergekit-community/L3.1-Athena-l2-8B
26
  - model: mergekit-community/L3.1-Athena-l-8B
27
  tokenizer_source: mergekit-community/L3.1-Athena-l3-8B
28
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - merge
5
  - mergekit
6
  - lazymergekit
7
+ language:
8
+ - en
9
+ base_model:
10
+ - rootxhacker/Apollo-exp-8B
11
+ - mergekit-community/L3.1-Athena-k-8B
12
+ - mergekit-community/L3.1-Athena-l2-8B
13
+ - mergekit-community/L3.1-Athena-l-8B
14
+ - mergekit-community/L3.1-Athena-l3-8B
15
+ pipeline_tag: text-generation
16
+ library_name: transformers
17
  ---
18
 
19
+ # 🧠 ZeroXClem-Llama-3.1-8B-Athena-Apollo-exp
20
 
21
+ ## Overview
22
 
23
+ **ZeroXClem-Llama-3.1-8B-Athena-Apollo-exp** is a powerful AI model built through **Model Stock merging** using [MergeKit](https://github.com/cg123/mergekit). It merges several of the most capable and nuanced Llama-3.1-based models available on **Hugging Face**, optimized for performance across **instruction-following, roleplay, logic, coding, and creative writing** tasks.
24
 
25
+ By fusing diverse fine-tuned architectures into a cohesive blended model, this creation delivers excellent generalist abilities while retaining specialized strengths.
26
+
27
+ ---
28
+
29
+ ## 🔧 Merge Details
30
+
31
+ - **Merge Method:** `model_stock`
32
+ - **Base Model:** [`mergekit-community/L3.1-Athena-l3-8B`](https://huggingface.co/mergekit-community/L3.1-Athena-l3-8B)
33
+ - **Dtype:** `bfloat16`
34
+ - **Tokenizer Source:** `mergekit-community/L3.1-Athena-l3-8B`
35
+
36
+ ---
37
+
38
+ ## 💡 Models Merged
39
+
40
+ The following models contribute to this powerful fusion:
41
+
42
+ - [`rootxhacker/Apollo-exp-8B`](https://huggingface.co/rootxhacker/Apollo-exp-8B) — A rich blend focused on alignment, DPO, and SFT instruction tuning across Llama-3.1 variants.
43
+ - [`mergekit-community/L3.1-Athena-k-8B`](https://huggingface.co/mergekit-community/L3.1-Athena-k-8B) — Roleplay and safety-aligned merge based on Meta's Llama-3.1 foundation.
44
+ - [`mergekit-community/L3.1-Athena-l2-8B`](https://huggingface.co/mergekit-community/L3.1-Athena-l2-8B) — LoRA-enhanced with long-context and creative capability merges.
45
+ - [`mergekit-community/L3.1-Athena-l-8B`](https://huggingface.co/mergekit-community/L3.1-Athena-l-8B) — Deeply infused with LoRA-based domain-specific models in logic, psychology, storytelling, and more.
46
+
47
+ ---
48
 
49
+ ## 🧪 Configuration
50
+
51
+ ```yaml
52
  name: ZeroXClem-Llama-3.1-8B-Athena-Apollo-exp
53
  base_model: mergekit-community/L3.1-Athena-l3-8B
54
  dtype: bfloat16
 
59
  - model: mergekit-community/L3.1-Athena-l2-8B
60
  - model: mergekit-community/L3.1-Athena-l-8B
61
  tokenizer_source: mergekit-community/L3.1-Athena-l3-8B
62
+ ```
63
+
64
+ ---
65
+
66
+ ## ✨ Features & Highlights
67
+
68
+ 🔹 **Instruction-Following Prowess** — Merged from Tulu-aligned and instruct-tuned models like Apollo-exp and Athena-k for high-quality, context-aware responses.
69
+
70
+ 🔹 **Immersive Roleplay & Personality** — Strong roleplay personas and emotional nuance thanks to Athena's diverse RP blends.
71
+
72
+ 🔹 **Creative & Structured Generation** — Support for creative writing, long-context novelization, and formal logic modeling from l2/l3 integrations.
73
+
74
+ 🔹 **Depth in Dialogue** — Enhanced ability to carry layered and philosophical conversation from Claude-style fine-tunes in Apollo-exp.
75
+
76
+ ---
77
+
78
+ ## 🎯 Use Cases
79
+
80
+ - **Conversational AI & Roleplay Bots**
81
+ - **Formal Reasoning & Chain-of-Thought Tasks**
82
+ - **Creative Writing & Storytelling Tools**
83
+ - **Coding Assistants**
84
+ - **Educational and Research Applications**
85
+
86
+ ---
87
+
88
+ ## 🛠️ Usage Instructions
89
+
90
+ ```python
91
+ from transformers import AutoModelForCausalLM, AutoTokenizer
92
+
93
+ model_name = "ZeroXClem/Llama-3.1-8B-Athena-Apollo-exp"
94
+
95
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
96
+ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype="auto", device_map="auto")
97
+
98
+ prompt = "Explain quantum entanglement like I'm 10 years old."
99
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
100
+ outputs = model.generate(**inputs, max_new_tokens=200)
101
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
102
+ ```
103
+
104
+
105
+ ### 🦙 Ollama Instructions
106
+
107
+ ```bash
108
+ ollama run hf.co/ZeroXClem/Llama-3.1-8B-Athena-Apollo-exp-Q4_K_M-GGUF
109
+ ```
110
+
111
+ ---
112
+
113
+ ## 🧭 Alignment & Ethics
114
+
115
+ ⚠️ **Unfiltered Output**: This model is uncensored and may generate content outside of alignment norms. Please implement your own **moderation layers** when deploying in production environments.
116
+
117
+ ⚠️ **Responsible Use**: Developers are encouraged to audit outputs and maintain ethical usage policies for downstream applications.
118
+
119
+ 📜 **License**: Usage governed by the [Meta Llama 3.1 Community License](https://huggingface.co/meta-llama/Llama-3.1-8B).
120
+
121
+ ---
122
+
123
+ ## 💌 Feedback & Contributions
124
+
125
+ We welcome your feedback, benchmarks, and improvements! Please open an issue or PR to contribute or tag us in your results and projects.
126
+
127
+ ---
128
+
129
+ **ZeroXClem Team | 2025**