Update README.md
Browse files
README.md
CHANGED
@@ -5,6 +5,123 @@ base_model:
|
|
5 |
- unsloth/Qwen2.5-1.5B-Instruct-unsloth-bnb-4bit
|
6 |
---
|
7 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
|
9 |
```
|
10 |
## License
|
|
|
5 |
- unsloth/Qwen2.5-1.5B-Instruct-unsloth-bnb-4bit
|
6 |
---
|
7 |
|
8 |
+
This is the LoRA adapter for Andy-4-micro, any use of this requires full adherence to the [Andy license](LICENSE)
|
9 |
+
|
10 |
+
# 🤏 Andy-4-micro ⛏️
|
11 |
+
|
12 |
+
**Andy-4-micro** is a compact and efficient model designed specifically for Minecraft gameplay via the Mindcraft framework. Built on the **Qwen2.5 1.5B** architecture, it delivers surprising capabilities given its size, offering strong reasoning and reliable task execution despite being only **21% the size** of the full Andy-4 model.
|
13 |
+
|
14 |
+
Andy-4-micro is licensed under the [Andy License 1.0], which requires derivatives to include the license terms and attribute the original author.
|
15 |
+
|
16 |
+
## 🧠 Overview
|
17 |
+
|
18 |
+
Andy-4-micro was originally an experiment—a testbed to see how much performance could be retained when compressing the Andy-4 training process into a significantly smaller base model. The result is a fast, efficient model with an excellent ability to play Minecraft and respond with logic and clarity, even under limited hardware constraints.
|
19 |
+
|
20 |
+
Despite its size, Andy-4-micro is able to:
|
21 |
+
|
22 |
+
- Understand and navigate Minecraft environments.
|
23 |
+
- Reason before responding, rather than simply reacting.
|
24 |
+
- Play autonomously or in collaboration with user input.
|
25 |
+
- Maintain high efficiency even on low-VRAM GPUs.
|
26 |
+
|
27 |
+
|
28 |
+
## 📊 Training Details
|
29 |
+
|
30 |
+
Andy-4-micro was trained on the same data as **Andy-4-base**, but with smaller architecture and unique training hyperparameters to optimize for compact performance.
|
31 |
+
|
32 |
+
- **Base Model**: `Qwen2.5 1.5B Instruct`
|
33 |
+
- **Total Training Epochs**: 4
|
34 |
+
- **Andy-4-base-2 dataset**: 2 epochs at `4e-5` learning rate
|
35 |
+
- **FT dataset**: 2 epochs at `8e-6` learning rate
|
36 |
+
- **Optimizer**: AdamW with cosine decay schedule
|
37 |
+
- **Quantization**: 4-bit (`bnb-4bit`)
|
38 |
+
|
39 |
+
This configuration allows Andy-4-micro to punch well above its weight class, especially when compared with other small models in the Minecraft AI space.
|
40 |
+
|
41 |
+
|
42 |
+
## 🔍 Differences from Andy-4-base
|
43 |
+
|
44 |
+
| Feature | Andy-4-base | Andy-4-micro |
|
45 |
+
|------------------------|----------------------------------|---------------------------------|
|
46 |
+
| Model Size | 8B | 1.5B |
|
47 |
+
| Base Architecture | DeepSeek-R1 Distill Llama 8B | Qwen2.5 1.5B |
|
48 |
+
| Training Duration | ~3 weeks | ~3 days |
|
49 |
+
| Usage Type | General Mindcraft agent base | Fast, efficient model for light gameplay |
|
50 |
+
| Reasoning | High | Medium-high |
|
51 |
+
| Hardware Requirements | 8+ | Very low (4-6GB VRAM works!) |
|
52 |
+
|
53 |
+
Andy-4-micro is not a replacement for Andy-4-base or the full Andy-4 model, but instead a supplementary option for users who prioritize speed and accessibility over scale.
|
54 |
+
|
55 |
+
---
|
56 |
+
|
57 |
+
## 🚀 Installation
|
58 |
+
|
59 |
+
You can install and run Andy-4-micro through Hugging Face and Ollama, following the same method as Andy-4-base. Here's how:
|
60 |
+
|
61 |
+
### 🧰 Hugging Face Quick Start
|
62 |
+
|
63 |
+
1. Go to the model page on Hugging Face.
|
64 |
+
2. Click on the **"Use this model"** dropdown.
|
65 |
+
3. Choose **Ollama** and select your quantization level:
|
66 |
+
|
67 |
+
| Quantization | Minimum VRAM Required |
|
68 |
+
|--------------|------------------------|
|
69 |
+
| F16 | 8+ GB |
|
70 |
+
| Q5_K_M | 6+ GB |
|
71 |
+
| Q4_K_M | 4-6 GB |
|
72 |
+
| Q3_K_M | 4 GB (low settings) |
|
73 |
+
| Q2_K | 2-4 GB (ultra-light) |
|
74 |
+
|
75 |
+
4. Follow the [Mindcraft Discord Guide](https://ptb.discord.com/channels/1303399789995626667/1347027684768878644/1347027684768878644) for cloud-based options if needed.
|
76 |
+
|
77 |
+
### 🛠️ Manual Installation via Modelfile
|
78 |
+
|
79 |
+
1. Download your chosen `.GGUF` quantization file and the `Modelfile`.
|
80 |
+
2. Open `Modelfile` and replace the `FROM` path with the exact path to your `.GGUF` file, e.g.:
|
81 |
+
|
82 |
+
```text
|
83 |
+
FROM C:\Users\you\Downloads\Andy-4-micro.Q3_K_M.gguf
|
84 |
+
```
|
85 |
+
|
86 |
+
3. In a terminal, navigate to the `Modelfile` directory and run:
|
87 |
+
|
88 |
+
```bash
|
89 |
+
ollama create andy-4-micro -f Modelfile
|
90 |
+
```
|
91 |
+
|
92 |
+
This will register the model as `andy-4-micro` for use with Ollama.
|
93 |
+
|
94 |
+
---
|
95 |
+
|
96 |
+
## 🧪 Experimental Nature
|
97 |
+
|
98 |
+
> **Note**: Andy-4-micro is an experimental release.
|
99 |
+
|
100 |
+
Although initial performance was overwhelmingly positive, Andy-4-micro was not originally designed for public deployment. However, due to strong community feedback and impressive early benchmarks, it has been released for broader use and further testing.
|
101 |
+
|
102 |
+
Expect updates, bugfixes, and potentially further fine-tuned versions based on user reports.
|
103 |
+
|
104 |
+
---
|
105 |
+
|
106 |
+
## ⚖️ License
|
107 |
+
|
108 |
+
This model is licensed under the **Andy License 1.0**.
|
109 |
+
Usage must include credit to the original creator, and derivatives must remain open and under the same license.
|
110 |
+
|
111 |
+
- Author: [Sweaterdog](https://huggingface.co/Sweaterdog)
|
112 |
+
- Acknowledgments: This model utilizes datasets and techniques developed in the Andy-4 ecosystem.
|
113 |
+
- License type: Derived from Apache 2.0 with added attribution & openness clauses.
|
114 |
+
|
115 |
+
|
116 |
+
## 📌 Final Notes
|
117 |
+
|
118 |
+
- Andy-4-micro is ideal for **lightweight deployments**, **educational use**, and **embedded inference**.
|
119 |
+
- It’s a great fit for users who want a responsive, low-resource model that still understands Minecraft deeply.
|
120 |
+
- For best results, use within the Mindcraft framework or alongside other Andy models.
|
121 |
+
|
122 |
+
We hope you enjoy exploring and building with **Andy-4-micro**!
|
123 |
+
|
124 |
+
> _"Big things come in small packages."_ 💡
|
125 |
|
126 |
```
|
127 |
## License
|