Initial upload of ProseFlow-v1-360M-Instruct-GGUF
Browse files- .gitattributes +3 -0
- ProseFlow-v1-360M-Instruct-Q4_K_M.gguf +3 -0
- ProseFlow-v1-360M-Instruct-Q8_0.gguf +3 -0
- ProseFlow-v1-360M-Instruct-imatrix.gguf +3 -0
- README.md +111 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
ProseFlow-v1-360M-Instruct-imatrix.gguf filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
ProseFlow-v1-360M-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
ProseFlow-v1-360M-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
ProseFlow-v1-360M-Instruct-Q4_K_M.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2e2cd2d6f724a562e6fff2fe19fa53d12a83ec91837bcb5ef287c85130ffff05
|
| 3 |
+
size 314829504
|
ProseFlow-v1-360M-Instruct-Q8_0.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:35e793fb78e0b805855b5d38a48868880ae83d1b8f91a1c777373a007532cf93
|
| 3 |
+
size 430643904
|
ProseFlow-v1-360M-Instruct-imatrix.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7df578bc333fe3c460aa41f09a69de270da99c7566b501e0af1cea1e264097aa
|
| 3 |
+
size 1099456
|
README.md
ADDED
|
@@ -0,0 +1,111 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: LSXPrime/ProseFlow-v1-360M-Instruct
|
| 3 |
+
base_model_relation: quantized
|
| 4 |
+
language:
|
| 5 |
+
- en
|
| 6 |
+
library_name: gguf
|
| 7 |
+
pipeline_tag: text-generation
|
| 8 |
+
license: apache-2.0
|
| 9 |
+
datasets:
|
| 10 |
+
- LSXPrime/ProseFlow-Actions-v1
|
| 11 |
+
tags:
|
| 12 |
+
- text-generation
|
| 13 |
+
- instruction
|
| 14 |
+
- proseflow
|
| 15 |
+
- unsloth
|
| 16 |
+
- smollm
|
| 17 |
+
- writing-assistant
|
| 18 |
+
---
|
| 19 |
+
|
| 20 |
+
# ProseFlow-v1-360M-Instruct
|
| 21 |
+
|
| 22 |
+
**ProseFlow-v1-360M-Instruct** is a lightweight, experimental instruction-tuned model created for
|
| 23 |
+
the [ProseFlow desktop application](https://github.com/LSXPrime/ProseFlow). This model is a fine-tune of HuggingFace's [
|
| 24 |
+
**SmolLM-360M-Instruct**](https://huggingface.co/HuggingFaceTB/SmolLM-360M-Instruct) and was created to explore the
|
| 25 |
+
capabilities of smaller language models on a diverse set of text-processing tasks.
|
| 26 |
+
|
| 27 |
+
The model was fine-tuned on the **[ProseFlow-Actions-v1
|
| 28 |
+
](https://huggingface.co/datasets/LSXPrime/ProseFlow-Actions-v1)** dataset.
|
| 29 |
+
|
| 30 |
+
**Note:** This model is provided for research and experimental purposes and low-resource devices. For the best user
|
| 31 |
+
experience in the ProseFlow application, the larger and more capable [
|
| 32 |
+
`ProseFlow-v1-1.5B-Instruct`](https://huggingface.co/LSXPrime/ProseFlow-v1-1.5B-Instruct) model is strongly recommended.
|
| 33 |
+
|
| 34 |
+
## Model Description
|
| 35 |
+
|
| 36 |
+
ProseFlow is a universal AI text processor that allows users to create and execute custom AI "Actions" on text in any
|
| 37 |
+
application. This model was an experiment to see if a ~360M parameter model could reliably perform the wide range of
|
| 38 |
+
tasks defined in the training dataset.
|
| 39 |
+
|
| 40 |
+
### Performance and Capabilities
|
| 41 |
+
|
| 42 |
+
Evaluations show that while this model is extremely fast and has very low resource requirements, its capabilities are
|
| 43 |
+
limited.
|
| 44 |
+
|
| 45 |
+
#### Strengths:
|
| 46 |
+
|
| 47 |
+
* **Extremely Lightweight:** Can run on devices with very limited RAM and computational power.
|
| 48 |
+
* **Strict Formatting Adherence (sometimes):** In some cases where it understands the task, it can follow rigid
|
| 49 |
+
formatting instructions (like creating a bulleted list) more strictly than its larger counterpart.
|
| 50 |
+
* **Simple Data Extraction:** It shows some capability in basic data extraction and formatting tasks, such as creating
|
| 51 |
+
Markdown tables or extracting contact information.
|
| 52 |
+
|
| 53 |
+
## Provided Files & Quantization Details
|
| 54 |
+
|
| 55 |
+
This repository provides multiple versions of the model, allowing users to choose the best balance of performance and
|
| 56 |
+
resource usage for their specific hardware. All quantized versions are provided in the GGUF format for broad
|
| 57 |
+
compatibility.
|
| 58 |
+
|
| 59 |
+
| File Name (Quantization) | VRAM Usage (Approx.) | Performance | Recommended Use Case |
|
| 60 |
+
|:-------------------------|:---------------------|:---------------------------------------------------|:--------------------------------------------|
|
| 61 |
+
| `Q8_0` | ~1 GB | **Best Overall.** Nearly identical to FP16. | **The recommended default for most users.** |
|
| 62 |
+
| `Q4_K_M` | ~900 MB | **Low Quality.** Noticeable degradation in nuance. | For maximum speed on low-power devices. |
|
| 63 |
+
|
| 64 |
+
**Note on Quantization:** To maintain the highest possible quality, the token embeddings and the final output layer were
|
| 65 |
+
kept at F16 precision. Additionally, an importance matrix was used for calibration during the quantization process. This
|
| 66 |
+
is why the quantized files are larger than what might typically be expected, as this method significantly improves their
|
| 67 |
+
performance and coherence.
|
| 68 |
+
|
| 69 |
+
#### Weaknesses & Limitations:
|
| 70 |
+
|
| 71 |
+
* **Poor Reasoning:** The model struggles significantly with tasks that require logical reasoning, inference, or
|
| 72 |
+
multi-step problem-solving. It often fails on word problems and logical puzzles.
|
| 73 |
+
* **Limited Creativity:** It is not effective at creative writing tasks like continuing a story or generating novel
|
| 74 |
+
content. Its outputs are often repetitive or nonsensical.
|
| 75 |
+
* **Instructional Failures:** The model frequently violates the "no extra text" rule by adding conversational chatter.
|
| 76 |
+
In many cases, it fails the task entirely and repeats the input verbatim.
|
| 77 |
+
* **Hallucination:** On some tasks (e.g., `To Paragraph`), the model hallucinates content completely unrelated to the
|
| 78 |
+
input.
|
| 79 |
+
* **Unreliable for Complex Tasks:** It is not suitable for complex tasks like code refactoring, bug finding, or drafting
|
| 80 |
+
professional business correspondence.
|
| 81 |
+
|
| 82 |
+
### Intended Use
|
| 83 |
+
|
| 84 |
+
This model is intended for **experimental use** and for users on **extremely resource-constrained systems** who are
|
| 85 |
+
willing to accept a significant trade-off in performance and reliability. It may be suitable for a very limited subset
|
| 86 |
+
of simple, repetitive text-formatting tasks.
|
| 87 |
+
|
| 88 |
+
It is designed to be used within the **ProseFlow desktop application**, but it is **not the recommended model for
|
| 89 |
+
general use**.
|
| 90 |
+
|
| 91 |
+
## How to Use in ProseFlow
|
| 92 |
+
|
| 93 |
+
1. [Download and install the ProseFlow application](https://github.com/LSXPrime/ProseFlow/releases).
|
| 94 |
+
2. Navigate to the **Providers -> Local Provider** tab.
|
| 95 |
+
3. Click "Manage Models..." and select the desired version of `ProseFlow-v1-360M-Instruct` from the "Available for
|
| 96 |
+
Download" list. **We recommend starting with `Q8_0`.**
|
| 97 |
+
4. Once downloaded, select it from the "My Models" list.
|
| 98 |
+
5. Set your "Primary Service Type" in ProseFlow to **Local**.
|
| 99 |
+
6. Be aware of the limitations described above when executing actions.
|
| 100 |
+
|
| 101 |
+
## Training Details
|
| 102 |
+
|
| 103 |
+
* **Base Model:** [HuggingFaceTB/SmolLM-360M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM-360M-Instruct)
|
| 104 |
+
* **Dataset:** [LSXPrime/ProseFlow-Actions-v1](https://huggingface.co/datasets/LSXPrime/ProseFlow-Actions-v1)
|
| 105 |
+
* **Fine-tuning Library:** [Unsloth](https://github.com/unslothai/unsloth)
|
| 106 |
+
* **Fine-tuning Method:** Supervised fine-tuning using LoRA on a dataset of structured instruction-input-output
|
| 107 |
+
triplets.
|
| 108 |
+
|
| 109 |
+
## License
|
| 110 |
+
|
| 111 |
+
This model is licensed under the [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
|